The Stages of Music Production: Pre-Production to Mastering
Music production is not a single act but a sequence of interdependent stages, each with its own craft, tools, and failure modes. This page maps those stages — from the earliest pre-production decisions through recording, editing, mixing, and mastering — and explains how choices at each phase constrain or enable everything that follows. Whether a track is being built in a bedroom studio or a commercial facility, the underlying structure of the process holds.
- Definition and scope
- Core mechanics or structure
- Causal relationships or drivers
- Classification boundaries
- Tradeoffs and tensions
- Common misconceptions
- Checklist or steps (non-advisory)
- Reference table or matrix
Definition and scope
The phrase "music production" describes the full chain of decisions and technical operations that transforms a musical idea into a finished, distributable recording. It spans at least 5 distinct stages — pre-production, recording, editing, mixing, and mastering — though professional practice sometimes subdivides these further (arrangement, sound design, stem mastering) or collapses them depending on the project's scale and budget.
The scope extends well beyond pressing "record." Pre-production alone can determine as much as 70–80% of a song's final quality before a single microphone is placed, according to the working principle widely endorsed by producers including Rick Rubin and documented in interviews archived by Sound on Sound magazine. Mastering, at the opposite end, is a narrower but legally significant process — streaming platforms including Spotify and Apple Music specify loudness normalization targets (Spotify targets –14 LUFS integrated, per Spotify's loudness normalization documentation) that masters must anticipate.
The process described here applies primarily to recorded music in the Western commercial tradition, though the same stage logic underpins music production across genres from hip-hop to orchestral film scoring.
Core mechanics or structure
Pre-Production
Pre-production is where a song is stress-tested before anyone spends studio time or money. Activities include finalizing the song structure, selecting keys and tempos, deciding on instrumentation, and creating reference mixes or demos. A tempo decision made here — say, 92 BPM versus 96 BPM — cascades into every subsequent stage, affecting quantization grids, delay time calculations, and even the feel of a live drum performance.
Recording
Recording captures audio performances into a digital audio workstation (DAW) or, less commonly in contemporary practice, analog tape. Multitrack recording allows instruments and voices to be captured on separate tracks, enabling independent editing later. Signal chain decisions — microphone type and placement, preamp gain, and analog-to-digital conversion — are locked in at this stage. A 24-bit/96 kHz recording captures significantly more headroom and frequency resolution than a 16-bit/44.1 kHz file, a distinction that matters during mixing.
Editing
Editing refines the raw recordings without altering tonal character. Tasks include comping (assembling the best sections of multiple takes), pitch correction using tools like Melodyne or Auto-Tune, time alignment, and noise reduction. Editing is where a performance becomes consistent; it is also where over-processing can strip a recording of its humanity in about three mouse clicks.
Mixing
Music mixing fundamentals occupy a stage where all recorded and programmed elements are balanced, spatialized, and processed into a cohesive stereo (or surround) field. The core tools are EQ, compression, reverb, delay, and routing. A professional mix engineer may spend 8–12 hours on a single track.
Mastering
Mastering is the final processing stage applied to the stereo mix (or stems, in stem mastering) to prepare it for distribution. It addresses broad tonal balance, loudness optimization, and format-specific encoding. Mastering does not fix mix problems — a point returned to in the misconceptions section below.
Causal relationships or drivers
The stages form a dependency chain: each upstream decision narrows the solution space downstream. A vocal recorded with excessive room noise cannot be fully rescued in editing; a mix with competing low-mid frequencies between the kick drum and bass guitar cannot be fully resolved in mastering.
Three factors drive how tightly these dependencies bind:
-
Signal-to-noise ratio at the source. Clean, well-gained recordings give editors more latitude. Recordings with noise floors above –60 dBFS start accumulating audible artifacts after heavy processing.
-
Arrangement density. Dense arrangements — 32 or more simultaneous tracks of similar frequency content — create mixing complexity that editing cannot anticipate. Pre-production decisions about instrumentation directly set this density.
-
Format target. A track destined for vinyl mastering has different loudness and dynamic constraints than one targeted at streaming platforms. Spotify's –14 LUFS target and Apple Music's –16 LUFS target (Apple's documentation on sound check) create ceiling pressures that flow backward into mixing decisions about dynamic range.
Sound design fundamentals and music arrangement decisions made in pre-production are the single highest-leverage points in this causal chain.
Classification boundaries
Not all production workflows follow the linear 5-stage model. Three alternative structures are common in professional practice:
Iterative production — common in electronic music and beat-making — collapses pre-production, recording, and editing into a continuous loop. A producer building a track in Ableton Live may record, edit, and arrange simultaneously across a session. This is documented extensively in production education resources like Berklee Online's course materials.
Remote/asynchronous production — increasingly standard in commercial work — distributes stages across locations and collaborators. A songwriter in Nashville might send stems to a mixer in Los Angeles and a mastering engineer in London, with each stage handled by a specialist who never meets the others.
Live-to-stereo capture — used for jazz, classical, and some folk recording — compresses recording and mixing into a single event, with mastering as the only post-capture stage.
The classification boundary that matters most practically: mixing and mastering are distinct disciplines. Submitting a rough mix to a mastering engineer in expectation that mastering will serve as mixing is a category error that professional mastering facilities like Sterling Sound explicitly address in their client documentation.
Tradeoffs and tensions
Speed versus quality in pre-production. Rushed pre-production saves days but frequently costs weeks in the back end. A song recorded in the wrong key for a vocalist will require either a re-record or pitch-shifted guitars — both expensive.
Loudness versus dynamics. The so-called "loudness wars" of the 1990s and 2000s — documented by researchers including Thomas Lund at TC Electronic — demonstrated that hyper-compressed masters score lower on listener preference tests despite being perceived as "more powerful" on first exposure. Streaming platform normalization has partially neutralized the competitive advantage of loud masters, but the tension persists in mastering decisions.
Specialization versus integration. A single producer handling all 5 stages moves faster and maintains vision coherence but brings fewer specialized ears to each discipline. A 3-person team — producer/recording engineer, mix engineer, mastering engineer — adds cost and communication overhead but typically raises the ceiling on each stage's quality.
Analog warmth versus digital precision. Analog hardware introduces even-order harmonic distortion that many engineers find pleasing; it also introduces noise floors, calibration drift, and maintenance costs. The professional recording studio versus home studio tradeoff is in part a negotiation between these two poles.
Common misconceptions
"Mastering will fix the mix." Mastering engineers operate on a 2-track stereo file or a small number of stems. They cannot isolate the snare drum, adjust the vocal level independently, or remove a muddy guitar part. Mastering addresses macro tonal balance, loudness, and format preparation — not the internal balance of the mix.
"More plugins equal better sound." Plugin accumulation is one of the most reliable ways to degrade a mix through phase issues, CPU-induced latency, and loss of transient clarity. The music production software and plugins landscape offers thousands of options; professional practice typically relies on a narrow, deeply understood toolkit.
"Louder is always better." Spotify, Apple Music, and YouTube all apply loudness normalization, meaning a master pushed to –6 LUFS will be turned down to the platform target, while a master at –14 LUFS will not. Excessive limiting to achieve loudness sacrifices dynamic range for no competitive gain on normalized platforms.
"Pre-production is optional for experienced musicians." Experience reduces but does not eliminate the cost of poor pre-production. Even experienced engineers note in production forums and interview archives (including Tape Op magazine) that unresolved arrangement problems discovered in mixing sessions are the leading cause of budget overruns.
For a structured introduction to the full ecosystem of production tools, the home studio setup guide and digital audio workstations explained pages cover the hardware and software layer that underlies every stage described here. The broader context of how all these moving parts fit together is covered at the musicproductionauthority.com home.
Checklist or steps (non-advisory)
The following represents the standard sequence of production stage milestones in a conventional commercial track workflow:
Pre-Production
- [ ] Song structure finalized (verse/chorus/bridge map with bar counts)
- [ ] Key and tempo confirmed and locked
- [ ] Instrumentation list complete
- [ ] Reference tracks selected (3–5 comparable commercial tracks)
- [ ] Demo or guide track recorded
Recording
- [ ] Signal chain tested and gain staged (peaks targeting –18 dBFS average)
- [ ] Sample rate and bit depth set (minimum 24-bit/44.1 kHz)
- [ ] Scratch tracks removed or clearly labeled
- [ ] All takes labeled by instrument, date, and take number
- [ ] Session backup confirmed on at least 2 separate drives
Editing
- [ ] Comping completed for all lead vocals and featured instruments
- [ ] Timing corrections applied and checked against original feel
- [ ] Pitch correction passes completed
- [ ] Breath and noise edits done
- [ ] Edited session archived separately from raw session
Mixing
- [ ] Gain staging checked at session open (mix bus not exceeding –6 dBFS before limiting)
- [ ] Low-end mono compatibility confirmed
- [ ] Stereo field checked on mono speaker (Auratone or similar)
- [ ] Revision passes completed (typically 2–3 rounds)
- [ ] Final mix exported at 24-bit/44.1 kHz (or 48 kHz for sync/film)
Mastering
- [ ] Mix file confirmed free of brickwall limiting or clipping
- [ ] Target loudness confirmed per platform (e.g., –14 LUFS for Spotify)
- [ ] ISRC codes assigned
- [ ] Final format exports completed (WAV for distribution, MP3 at 320 kbps for promo)
Reference table or matrix
| Stage | Primary Inputs | Primary Outputs | Key Tools | Typical Duration (single track) |
|---|---|---|---|---|
| Pre-Production | Song idea, arrangement concept | Demo, tempo/key map, reference tracks | DAW, instrument, pencil | 2–8 hours |
| Recording | Demo, musicians, signal chain | Raw multitrack session files | Microphones, preamps, DAW, audio interface | 2–16 hours |
| Editing | Raw session | Comped, aligned, pitch-corrected session | Melodyne, DAW elastic audio, noise gates | 1–6 hours |
| Mixing | Edited session | Stereo mix file (typically WAV 24-bit) | EQ, compression, reverb, delay, DAW | 4–12 hours |
| Mastering | Stereo mix (or stems) | Distribution-ready audio files | Mastering EQ, multiband compressor, limiter, metering | 1–3 hours |