🎼Electronic Music Composition Unit 1 – Electronic Music & Sound Fundamentals
Electronic music composition blends technology and creativity, using electronic devices and software to craft unique sounds. This unit explores the fundamentals of sound, from basic physics to digital audio concepts, providing a foundation for understanding how electronic music is created.
Synthesis techniques, MIDI, and audio effects are key tools in electronic music production. The unit also covers essential DAW skills, composition techniques, and the art of sound design, equipping students with the knowledge to bring their musical ideas to life.
Electronic music composition involves creating music using electronic devices, software, and techniques
Sound is a vibration that travels through a medium (air, water, solid materials) and can be perceived by the human ear
Waveforms represent the shape and characteristics of a sound wave, with common types including sine, square, sawtooth, and triangle waves
Frequency refers to the number of cycles a sound wave completes per second, measured in Hertz (Hz)
Amplitude is the strength or intensity of a sound wave, often perceived as loudness
Timbre is the unique character or quality of a sound that distinguishes it from other sounds with the same pitch and volume
Envelope describes how a sound changes over time, typically divided into attack, decay, sustain, and release (ADSR) phases
Oscillators generate basic waveforms and serve as the foundation for sound synthesis
Sound Physics and Waveforms
Sound waves are longitudinal waves that cause particles in a medium to compress and expand, creating areas of high and low pressure
The speed of sound varies depending on the medium, with sound traveling faster in solids than in liquids or gases
Wavelength is the distance between two corresponding points on a wave, such as two consecutive peaks or troughs
The relationship between frequency and wavelength is inversely proportional: higher frequencies have shorter wavelengths, while lower frequencies have longer wavelengths
Harmonics are integer multiples of a fundamental frequency that contribute to the overall timbre of a sound
Overtones are any frequencies above the fundamental frequency, including both harmonics and inharmonic frequencies
Phase refers to the position of a point on a waveform cycle relative to the start of the cycle, measured in degrees or radians
Constructive and destructive interference occur when sound waves interact, resulting in increased or decreased amplitude, respectively
Digital Audio Basics
Analog-to-digital conversion (ADC) is the process of converting continuous analog audio signals into discrete digital data
Sample rate is the number of times per second that an analog signal is measured and converted into a digital value, typically expressed in Hz or kHz
Higher sample rates (44.1 kHz, 48 kHz, 96 kHz) capture more detail and higher frequencies but result in larger file sizes
Bit depth is the number of bits used to represent each sample, determining the dynamic range and signal-to-noise ratio of the digital audio
Common bit depths include 16-bit (CD quality), 24-bit (studio recording), and 32-bit (high-resolution audio)
Nyquist theorem states that the sampling rate must be at least twice the highest frequency in the analog signal to accurately represent it in the digital domain
Aliasing occurs when the sampling rate is too low to capture high frequencies, resulting in these frequencies being misrepresented as lower frequencies
Dithering is the process of adding low-level noise to digital audio to minimize quantization errors and improve the perceived resolution of the signal
Digital-to-analog conversion (DAC) converts digital audio data back into a continuous analog signal for playback through speakers or headphones
Synthesis Techniques
Additive synthesis creates complex sounds by combining simple waveforms (sine waves) at different frequencies and amplitudes
Subtractive synthesis starts with a harmonically rich waveform (square, sawtooth) and filters out unwanted frequencies to shape the sound
Wavetable synthesis uses pre-recorded or generated waveforms stored in a table, allowing for smooth transitions between different waveforms
Frequency modulation (FM) synthesis creates complex timbres by using one oscillator (the modulator) to modulate the frequency of another oscillator (the carrier)
Granular synthesis divides a sound into short segments called grains, which can be manipulated and recombined to create new textures and timbres
Physical modeling synthesis simulates the physical properties and behavior of real-world instruments using mathematical algorithms
Sampling involves recording and manipulating real-world sounds, which can be played back and processed in various ways
Synthesis parameters, such as oscillator pitch, filter cutoff frequency, and envelope settings, can be modulated to create dynamic and evolving sounds
MIDI and Digital Instruments
MIDI (Musical Instrument Digital Interface) is a protocol for communicating musical performance data between electronic devices
MIDI messages include note on/off, pitch, velocity, and control change information, but do not contain actual audio data
MIDI sequencing allows for the recording, editing, and playback of MIDI performances, enabling precise control over timing and expression
Virtual instruments are software-based synthesizers or samplers that respond to MIDI input and generate audio output
MIDI controllers, such as keyboards, drum pads, and wind controllers, provide tactile control over MIDI parameters and can be used to perform virtual instruments
MIDI mapping allows for the assignment of MIDI messages to specific parameters in software or hardware, enabling customized control over sound and performance
MIDI clock is a timing signal that synchronizes tempo-based devices, such as sequencers, drum machines, and effects processors
MIDI automation records changes to MIDI parameters over time, allowing for dynamic and evolving performances
Audio Effects and Processing
Equalization (EQ) is the process of adjusting the balance of frequency components in an audio signal, used for shaping the tonal character of a sound
Compression reduces the dynamic range of an audio signal by attenuating loud parts and/or amplifying quiet parts, creating a more consistent level
Reverb simulates the natural reverberation of a physical space, adding a sense of depth and space to a sound
Delay creates a repeating echo effect by playing back a delayed copy of the original signal
Chorus creates a thickening effect by combining the original signal with slightly detuned and delayed copies of itself
Flanger creates a sweeping, metallic effect by mixing the original signal with a delayed copy and varying the delay time
Distortion adds harmonic overtones and non-linear effects to a signal, creating a gritty or aggressive sound
Modulation effects, such as phaser, tremolo, and vibrato, create movement and interest by varying parameters like phase, amplitude, or pitch over time
DAW Essentials
A Digital Audio Workstation (DAW) is a software environment used for recording, editing, and producing audio and MIDI data
Multitrack recording allows for the simultaneous or sequential recording of multiple audio or MIDI tracks, which can be independently edited and processed
Mixing is the process of balancing and blending individual tracks, applying effects, and creating a cohesive stereo or surround sound image
Automation in a DAW refers to the recording and playback of changes to parameters over time, such as volume, panning, or effect settings
Audio editing tools include cut, copy, paste, fade, and crossfade functions, enabling precise manipulation of audio regions
MIDI editing tools allow for the quantization, transposition, and velocity adjustment of MIDI notes and events
Virtual instruments and effects can be loaded as plug-ins within a DAW, expanding the creative possibilities for sound design and processing
Rendering or exporting a project creates a final audio file that can be distributed or further processed outside the DAW environment
Composition Techniques for Electronic Music
Layering involves combining multiple sounds or elements to create a rich and complex texture, often using complementary or contrasting timbres
Sequencing is the process of arranging and programming musical events (notes, chords, patterns) over time, typically using a MIDI sequencer or DAW
Sampling and looping techniques involve capturing and repeating audio segments to create rhythmic or melodic elements, often used in genres like hip-hop and electronic dance music
Sound design is the process of creating and manipulating sounds to achieve a desired aesthetic or emotional effect, often using synthesis, sampling, and effects processing
Arrangement refers to the structural organization of a composition, including the introduction, development, and variation of musical ideas over time
Modulation techniques, such as key changes, metric modulation, or timbral modulation, can add interest and variety to a composition
Generative and algorithmic composition techniques use mathematical models, rules, or chance operations to create musical structures or content
Collaboration with other artists, musicians, or producers can bring new perspectives, skills, and ideas to the compositional process