SECTION 1: Interpreting the Scene

I was meticulous (borderline neurotic) about dialing-in my setup: proper placement of speakers, multiple subwoofers, bass traps, acoustic treatments on walls and ceiling. Check. I tamed the room modes with parametric EQ on my Xilica XD-4080 and got the blend spot-on for subs and mains. Check. Shot one last plot with Room EQ Wizard and everything looked awesome. I was super thrilled and stayed happy.

That’s precisely what the tools are designed to do: reveal real problems, show whether a tweak is moving things in a useful direction, and get the system into a zone where the real listening can begin.

But I just couldn’t leave well enough alone…

Like most of us, I tried to move the bar higher. That 1/6th octave ‘bump and hump’ I saw in the REW plot was the little devil on my shoulder, just daring me to ‘fix’ it. I eventually succumbed to the darkside hobby, pulled out the calibrated mic and dialed-in a couple narrow parametric EQ adjustments to make everything look ‘right’ on my computer screen. That hump and bump was electronically beaten into submission by science, and the sound was excellent.

For a while…

After a week or so, I felt uneasy while listening… something was ‘off’. I defaulted to my normal routine: go back and undo the last tweak. I played the first track and quickly knew something changed for the better, but there was no obvious ‘thing’ I could grab onto as to ‘why’. Over weeks of listening, I'd go back and dial-in some EQ to make that plot look better and each time, I couldn’t detect exactly what was wrong… it just didn’t sound ‘right’.

Measuring ‘better’ in this instance didn’t mean it sounded better.

I shared this experience with a recording engineer friend, and we eventually turned to the subject of mics. With tongue firmly in cheek, I asked if he could simply EQ a ‘flat’ mic into performing like the legendary Neumann U47. His response — after a bit of a groan and slow shake of his head — was that you can get a mic to look like the U47 frequency response, but that doesn’t make it a U47. My friend’s engineering wisdom on that topic ended with: “measurements don’t always track the mojo.”

When our conversation turned into a geek session over music we both grew up on, my drumming heroes were on deck: John Bonham & Ian Paice, as well as the work of Glyn and Andy Johns – both pioneers in the recording world, especially as it relates to recording drums.

I brought up Glyn Johns, and he also knew the legendary story of how his brother Andy was able to get that huge sound on Led Zep’s When the Levee Breaks. This wasn’t just a few mics put in front of his kit in a studio. You’re hearing the character of a stone-built 18th century mansion (Headley Grange) with Bonham and his kit positioned at the bottom of a three story stairwell. Two heavily compressed Beyerdynamic M160 mics were positioned high above 'Bonzo' in the stairwell to capture the reverb. Then a Helios console, a Binson Echorec, and tweaked tape speed helped frame the scene. The snare wasn't just an ordinary Luddy 402: it transformed into an instrument now worthy of being struck by a drumming God. The 26" bass drum became so thunderous, Thor himself might consider another line of work.

This sound wasn’t simply ‘captured’ and documented. It was crafted.

Nobody in that room thought these choices were cheating. It was craftsmanship… later recognized as pure genius. Singers may choose one mic for more bite, or another for more air. Guitarists know when one mic is wrong and another one is better for that guitar, amp or song. Engineers choose ribbons, condensers, angles, preamps, echoes, tape speed, and overload points because the scene capture is not clerical work. It’s craft under pressure... with a tired back, a ticking clock, and musicians waiting on the talkback.

What we ultimately get to hear are the cumulative decisions that are rendered onto tape or into digits. The engineer frames the scene.

The rub is when you try to pin down what exactly the ‘accurate’ sound was. What version of the scene is correct? The musicians were the only people with a direct window to the performance. I should say… multiple windows: the drummer was inside the boundary of his kit, the singer heard the room through chest and jaw, the guitarist turned toward the amp instead of the room, the bass player locked with the drummer and feeling the note arrive in the fingers before anyone else. The musicians in that room were not standing in one shared acoustic truth. Then there’s the engineer: they get their own special ‘window’ of truth — their monitors, room, experience, mood on that unique day. It happened once. Then it was gone.

That was the human side. Each person collected memories that only they can own.

The window we get to hear is captured not by a human brain, but by a device. The device wasn’t 7th row center at Royal Albert Hall. A vocal mic may be close enough to catch spit on a hard consonant; a kick drum mic may live a foot from the resonant head where no human listener would dare stand. This is not a privileged seat. It’s a tool position, chosen precisely because it hears differently. Classical listeners will note that a Brahms symphony could have a live ‘reference’, but which one — Row 12? The conductor’s? The first oboist’s? The recording engineer’s mic position, which no audience member occupied?

Even in the most stable case we can imagine, the ‘ultimate reference’ dissolves the moment you try to pin it down. Every listener in the seats heard a slightly different acoustic event, and every memory of it kept moving after the lights came up and the venue emptied.

The ‘reference’ doesn’t hold still.

Of course, there is always the recording itself — fixed, final, and indifferent to who was in the room. This assumption deserves a closer look.

SECTION 2: Translating the Artifact

Geoff Emerick didn’t approach Beatles sessions like a flat-transfer exercise, and a Shure SM57 on a snare certainly isn’t a faithful ‘witness’. Bonham under M160s in that Headley Grange stairwell wasn’t clerical, and neither was Sinatra through a U47 mic.

Recording engineers intentionally select ribbons, condensers, preamps, overload points, console paths, etc… as part of the craft. The sound is being shaped before anybody at home gets a vote. Source coloration isn’t the failure case in professional recording. It’s the craft.

Even inside one mic manufacturer’s own family, the voicing can shift: Peus (1997) measured the same Neumann capsule across circuit generations and found the response shifted. Same capsule family. Different result.

Engineers who use two microphones that are spec’d similarly on paper but hear one clear the window, while the other casts a haze on the glass know the feeling. In Sennheiser’s own investigation early in the digital era, engineers complained about exactly that kind of transparency loss. CD-level ‘cleanliness’ unmasked what tape and vinyl had been forgiving for decades. The usual THD number wasn’t picking this up, and the mic capsule’s own distortion figures weren’t either.

When Sennheiser dug deeper with a two-tone test, the culprit was found in the mic capsule itself: the diaphragm didn’t move symmetrically into and out of its air gap. The subsequent push-pull MKH mic design attacked this problem by making the air-gap conditions symmetrical. German engineering for the ‘win’ — but the 'mother' of this win was the engineers who heard what the measurements missed.

One of the most consequential microphones in history is the Neumann M50. The three-microphone recording technique known as The Decca Tree grew around it and defined classical recording for decades. Engineers routinely refer to the M50’s depth, image stability, and spatial realism as if they’re describing a perspective, not a frequency curve.

Because they are.

EQ can copy the M50’s on-axis balance. It cannot copy how it weighted the source. That choice happened in air. It happened before electricity. The mojo the recording engineer hears is real. It just wasn’t found in the published spec sheet.

My dear friend Brian Kelly — a retired jazz guitar player who was doing session work at the age of 16 in 1960’s Cincinnati, and had the ear of major record label execs — recently told me what it was really like inside the recording industry… the part of the sausage getting made we don’t usually hear about. The engineers who choose the mic, the distance, the console, the compression, etc… get paid to deliver a record that works outside of their control room: Cars. Clubs. Headphones. Giant floorstanders in tuned rooms or cheap boomboxes in ugly ones.

This doesn’t cheapen the artifact... it helps define it. The engineer’s high-wire act means the record has to survive real world usage — not just our purpose-built listening rooms and five or six-figure systems. That’s the reality of the world Brian saw.

From the other side of the glass, my friend Christos Tsantilis (multi-platinum mixing/mastering engineer and acoustician) told me about the multiple layers of editing that are embedded in the process, starting with the first engineer who captures the audio:

"What they’re hearing during the mix isn’t even the final version that gets released. The mastering engineer can reshape the record — sometimes subtly, sometimes significantly. So the final master is faithful to the mix, but it’s also a refinement — and often a different experience.”

Mastering legend Bob Ludwig added another facet to the mastering process when he explained the evolution of Avenged Sevenfold’s The Stage. He cut a dynamic master aligned with his usual approach, and the band signed-off. A few days later, the band got nervous it wouldn’t hold up against the loudness of their peers — so Ludwig produced a compromise version. After some back-and-forth, the band circled back and chose the more dynamic original for release. Same engineer, same album, days apart — and the ‘right’ call shifted more than once.

A record is an artistic statement, but it is also a commercial product built to survive translation, and ultimately… sell. What leaves the building is not the untouched event. It is the strongest available artifact of the reference made at that moment despite the compromises of that era.

Time. That’s another factor. Do you recall that original vinyl pressing you owned, or the 180 gram reissue, ‘87 CD, 2003 remaster, or the 2015 hi-res issue that quietly replaced it in your library? If the original master was a fixed reference, should sanctioned revisions be viewed as ‘vandalism’? Should we recalibrate our systems to accommodate the new master?

A great example of the shifting ‘reference’ can be found during the making of the Eagles Hotel California for the DCC Gold Disc series. Steve Hoffman — widely regarded as one of the best mastering engineers ever — received the master tapes, but felt the sound was muffled and bass-heavy. He pulled the safety copies — identical. It wasn’t bad tape. It was the way the album had been originally mixed.

He nearly walked away from the project.

Hoffman then did something that would reverberate through the decades: he tracked down a similar pair of the JBL monitors the album had originally been mixed with. He then set them up in the middle of the room without any bass reinforcement, and… that same tape was now workable. The mastering decisions he made followed entirely from which speakers he was listening through, which included a 14 dB cut at 100 cycles (Hz). A measurement didn’t point to those JBLs — it was his ears, experience, and judgment.

As he put it, he had two choices: sell a pair of JBL monitors with every gold CD or… master it so it sounded ‘right’ on everyone’s stereo. He chose the latter. The tape didn’t change. He changed his perspective with the equipment. Had he never found those JBLs, a different Hotel California would have shipped.

What Hoffman ran into wasn’t a defective tape. The tape held the sum of all the choices made to that point — every mic choice, level tweak, compromise, and decision made through the locked master. But what it didn’t hold — what it couldn’t hold — was how it would be heard once it left the studio. That requires equipment, speakers, and a space where someone actually listens.

I think of the master tape as a fixed layer at the center of a longer chain, similar to a ‘digital negative’. The negative is simply bits on a hard drive or memory card... until it’s processed.

The artifact is fixed. The rendering engine gives it ‘life’.

The recording industry has known this for a long time. Standards bodies wrote specs for what a control room should sound like, acousticians designed the rooms, and studios spared no expense to build them. But there was a problem: rooms built to the same spec still sounded different to the engineers who worked inside them. Despite the recording industry’s best efforts, the same recorded artifact leaned differently in different rooms...

SECTION 3: The Renderer and the Body

The Room’s Lean

Let’s imagine the recording engineer at the final mix, totally unconstrained. No commercial motivation, nobody in their ear about ‘less bass’, or ‘more cowbell’… Just mixing for the best possible sound for their ears, using those monitors, amp and electronics on that specific day. The best possible scenario…

Now… change only the room. Different walls, different ceiling, different floor. Play exactly the same recording. Same speakers, same electronics, same listener... Does it sound the same or will it render differently?

Most of us know what an aggressively treated, over-damped room sounds like: the speakers can measure clean and technically 'correct', but the music veers toward ‘sterilized’ and mechanical. The soundstage can become narrower, with less depth. But this doesn’t necessarily mean a dry room is ‘wrong’. A space built for analysis can help recording engineers hear balances, edits, and low-level problems that matter to the work. But there is a limit: the recording industry has moved away from overly dead studios, after musicians routinely complained their instruments didn’t sound or feel right.

It wasn’t just musicians.

Sound reproduction expert Floyd Toole has highlighted work that recording engineers preferred a strong, direct sound field for mixing but preferred more reflections when listening at home. Inside the profession, Tervo and colleagues (2014) found a related split: mixing engineers leaned toward drier, higher-clarity environments, while mastering engineers leaned toward more reverberant ones. Even the 'dry' rooms in Tervo’s study didn’t cross over into a full-blown ‘anechoic’ state.

The chair that helps you work on a record is not automatically the chair that helps you live with it.

Bringing this into a more familiar world: in addition to mixing and mastering for platinum level artists, my engineer friend Christos also designs state-of-the-art mix rooms/studios. His recent project was for Rock & Roll Hall of Fame artist Steve Miller. When comparing ‘studio sound’ to what is preferred by most listeners in their own rooms, Christos’ response was:

Most Hi-Fi systems introduce a low-end lift — often somewhere between 1 to 6 dB, below 100 Hz — which adds weight and excitement. In contrast, mastering rooms are typically tuned to be flat, or at most have about a 1 dB rise in that same region.”

If the ‘perfect’ mix only sounds perfect in that specific room, with that specific equipment, by that specific engineer on that day… don't we end up with an audiophile version of the Texas Sharpshooter Fallacy — drawing a ‘target’ around holes... after the fact?

Glenn Meadows — who’s mastered over 1,300 gold and platinum records at Masterfonics — has said:

“I think that mastering is, and always has been, the real bridge between the pro audio industry and the hi-fi industry. We’re the ones that have to take this stuff that sounds hopefully good or great on a big professional monitor system and make sure it also translates well to the home systems.”

Gear Pressure

If we acknowledge the room leans on the signal, what’s the gear’s fingerprint on it? I love measurements, and use them every day for work… but like an xray, blueprint, or dyno on a car… they rarely offer the complete picture.

Noise matters. Frequency response matters. Stability, clipping behavior, output capability, linearity into defined conditions... all of this absolutely matters. But is testing a simple signal through a single component into a static load going to foretell what happens when a real music signal with full harmonic complexity moves through a chain of reactive components?

Audio Precision said it cleanly in a 2010 technical note: if you neglect reactive-load testing, you’re left speculating about actual use. Renardson made a similar point decades earlier: swap out the simple 8 ohm resistor ‘load’ you test an amp with, and put in a real loudspeaker. He wrote that THD with loudspeaker loads is typically about 10 dB higher, citing earlier work.

Of course this made sense when I first read it, because a speaker’s impedance (AC’s version of resistance) can swing wildly with frequency. It stores and returns energy back into the amp (back-EMF), and generally behaves nothing like the well-mannered fixed resistor the spec sheet was tested into.

I like to think of it like this: a manufacturer tests your car’s 0–60 time on a perfectly flat, dry, climate-controlled track — no traffic, no hills, no weather, no potholes. The car behaves beautifully. Take it on your actual commute: stop-and-go traffic, a steep hill, a gravel patch, a tailwind on the highway, or even… off-road? The engine has to constantly adapt to changing conditions, and now you notice behavior the spec sheet never hinted at. The test track wasn’t lying — the test just never challenged the car the way the real world does. A speaker is your commute. A resistor is the test track.

This dynamic doesn’t require an active source > load chain to show up. It also lives inside individual, passive components. Bateman (2002) measured a signal capacitor under real-world operating conditions, with signal running and DC bias applied. The twist: he used real-time measurement hardware instead of the conventional averaged approach. The distortion was there, but it wasn’t constant. It appeared and disappeared at irregular intervals that were intertwined with the signal itself. The conventional FFT average smoothed all of that into a nice smooth line.

The spec sheet wasn’t wrong — it just couldn’t properly weight what appeared between samples. The standard methods were still telling the truth about steady-state behavior. They just weren’t screening the whole 'movie'. Different tool. Different slice of reality.

When these distortion artifacts are injected on top of the harmonics and decay, the ‘mortar between the bricks’ is corrupted. A cymbal can keep its top-end energy but still lose the harmonic richness that gives it a unique personality. A snare or tom can keep its attack, but also feel like the overtones were bleached-out. This is the sort of thing that made me wonder: how do we define what ‘clean’ and ‘transparent’ really mean?

I learned the answer was always right in front of me: time. Trained normal-hearing listeners are capable of discriminating left-right timing differences below 10 microseconds, as Thavam and Dietz (2019) measured in spatial hearing. This tiny scale helped me gain new perspective on our sensitivity to timing differences, and made me wonder just how this might map to what we try to replicate in our hobby: creating a sense of 3-dimensional space that sounds real. Stuart and Craven (2019) mapped the question further into the audio chain: they modeled eight stages chained together and showed the accumulated timing errors can blow past their own 10-microsecond target without any single stage taking the blame. Time behavior stacks.

This brings me back to that ‘hump and bump’ I chased with my Xilica’s parametric EQ. Each PEQ move was supposed to clean up the frequency response, and the Room EQ Wizard trace beautifully confirmed the ‘fix’. What I didn’t understand at the time was that a filter is still a filter: magnitude, phase, and group delay travel together. I had elevated that frequency plot as if it were telling the whole story. My ears found the trap before I had an explanation for it. Later, it was Radlović and Kennedy’s work (2000) that pointed towards a possible ‘why’: a correction can reduce the visible magnitude error in the SPL graphs I was using, while also disrupting the timing behavior. After many rounds of beating my head against the wall chasing that ‘beautiful’ looking frequency graph, I learned to stop treating large, narrow corrections above the bass range as ‘free money’. The plot indicated ‘better’, but listening told me ‘worse’.

The tool wasn’t broken. It was just the wrong one for this job.

My listening room echoed this experience, but at a different scale. In 2011 I spent a good chunk of my summer doing what any slightly ‘off’ audiophile might call ‘fun’: armed with a calibrated mic and Room EQ Wizard, I measured every conceivable speaker placement option, xover settings, slopes, etc… and then walked two 1,000 watt subwoofers around the perimeter of my listening room inches at a time.

For the sub placement, I paid special attention to pesky nulls where no amount of EQ would correct the problem. What I didn’t expect was how differently my room presented these nulls. Left wall: big suck-out (null). Same spot on right wall: smooth as butter. After some brief head scratching, I figured it out… The left side wall and front wall are framed next to a poured concrete foundation. The right and back walls are freestanding. Each construction type produced measurably different ringing that interacted with my room’s modes. The conventional wisdom of ‘put your subs in the corners’ or ‘keep them symmetrical in the room’ didn’t square with what the measurements alerted me to. The optimal solution was asymmetric placement to work with the room and the materials it was made with, not fight against it. The measurement didn’t fail. It flagged something was ‘off’, and it was up to me to figure it out.

Feel the Vibe?

After my new speakers arrived, Bill Dudleston (Legacy Audio) was over helping me get things set up. During this session, Bill pointed out something that stuck with me: too much bass in a specific band doesn’t just color the sound — it can make the room physically pump. Many of us are familiar with the ‘one-note bass’ you can get with some subs or poor speaker placement...

This was different. I wasn’t just hearing this problem. My body could actually feel the room pumping.

I dig a nice headphone rig. But… it’s a very different experience from my main rig. For me, I’ve always felt like the body ‘listens’ too. But it wasn’t until I had a bad case of sinusitis, and I was taking a shower, when I spotted something interesting. As the shower water hit my head, I could hear it — even when I plugged my ears with my fingers. “Hmmm… this just might mean something.” This question became another audio related 'itch' that I needed to scratch...

When you think about it, it makes sense: our ears aren’t just hanging out there, isolated in free air like a microphone. Our ear’s ‘guts’ are mechanically connected to the rest of our body via fluid, tissue and bones… all good mechanical conductors of vibration.

Body-surface vibration under airborne low-frequency sound from loudspeakers was measured by Takahashi and colleagues (1999) and they found that chest and abdomen vibration increased with SPL in the tested range. The body behaves as a mechanical transducer — with the chest shown to be affected more than one's abdomen. This response also varies between body types: different build, different response.

This explains how the tactile punch of a bass drum or vibe from the bass player changes how I feel the groove. In the very best systems I’ve heard — you don’t just hear bass... You wear it. You feel it in the sternum, the floor through your feet, the chair back, the little tightening in the gut when kick drum and bass line lock together and the whole track starts pushing air with purpose.

That’s bass you feel and hear at the same time. If you’re a headphone aficionado, you’re probably all too aware of the arguments that "headphones don’t let you feel the bass", as well as the counterargument that it’s not needed for musical content. I enjoy both, so I usually kept my head down during those debates. But I did wonder… is body-felt bass really just for movie booms and rap thumps? My body said otherwise... and I eventually found science backed me up. Cameron and colleagues (2022) hit a live show with inaudible sub-bass: dancing jumped 11.8%. That’s pure body reflex kicking in. Siedenburg’s team (2024) found body-coupled bass cranks engagement far more than ‘liking’. You can’t name it, but it activates and engages you. No gimmick. Your chest keeps score, before the adjectives arrive.

Off-Rack or Tailored?

I’ve attended audio shows and have listened to the exact same track, in the same seat as some well-known gear reviewers. The opinion they came away with was sometimes miles apart from my assessment. Many of us have also listened to the same track on the same system at audiophile buddies’ homes and came away with a differing impression.

We often see this during the various audio shows, where the ‘best room’ or ‘best of show’ never seems to settle onto a single, or even consensus ‘best’. Wade into the audio forums and the spread of opinions is even wider than polling what the best pizza joint is in town.

This isn’t just mood, training, or stubbornness. Reviewers usually know their stuff. Our audiophile buddies usually have good ears too. So what’s at play here? Part of the answer was closer than I thought — literally attached to my head…

Our outer ear shapes every sound before it reaches our eardrums — and that shape is as unique as your fingerprint. Middlebrooks (1999) measured it: the sound shaping is mainly in the upper frequencies where your brain gets its directional and spatial cues — and it’s not something you can just dial in or out with a tone knob or EQ. Wenzel and colleagues (1993) found the practical fallout when researchers simulated three-dimensional sound using a ‘generic ear’ model. Listeners actually lost track of whether sounds were in front or behind, above or below — and each person got confused in their own way. The ‘generic ear’ might be an interesting lab creation, but it’s just not how people hear…

Going another layer deeper, the ear continues to show it's not 'off-the-rack'. There is a small opening deep in our ear’s ‘plumbing’ (the helicotrema) that acts like a pressure relief valve, and affects how we perceive low frequency sound. The interesting part is the size of this opening varies from person to person. This means one person’s "Prodigious bass!" assessment can be another’s "Crank that sub up, dude…"

Cherri and colleagues (2024) found that listeners who all passed a ‘standard’ hearing test could vary widely when it came to finer details like separating a voice from background noise, or locating a sound in space. A normal audiogram is like passing a driver’s vision test: you read the eye chart fine, but it says nothing about spotting a deer alongside the road or judging gaps in traffic. It's a ‘normal’ threshold label, not a perceptual fingerprint.

I was reminded of this when my local audiophile friend ‘Dave’ commented that he listened to drums differently after I walked through what I heard (as a drummer) in a track. These were cues that most drummers would reflexively listen for, but won't hit the radar of most non-drummers. I’m sure a pianist or cellist would absorb instruments differently than I do, as well.

Mixing engineer Jack Joseph Puig made a keen observation that really nails it:

I feel everybody has what I call a ‘preset’ — something they default to. The preset is developed by the synthesis of all your experiences as you’ve grown up listening to music.”

This explains a scene most of us know very well: experienced listeners sitting in the same room, same system, same sweet spot, same music — one is looking for an off-ramp, while the other gets lost in the music. Part of the rendering chain now lives in a chunk of cartilage, bone, and a few centimeters of very non-generic geometry. Our ear is the last piece of hardware — and it’s truly unique for each of us.

SECTION 4: Push Me… Pull Me

By the time I realized what was happening, my body was already there. Eyes closed, foot tapping, and lost in the music…

I’d spent days in ‘engineer mode’ for the listening session with some local audiophile buddies. Subwoofer tweaks, crossover tweaks, a new battery-operated network gadget... all the arrows in my quiver. Knowing the crew’s music preferences, I scripted the tracks like I was calling plays for the first series of a Packers game. The session started, and as we moved through tracks, the comments started coming about soundstaging, bass, etc...

Then something special happened: Stevie Ray Vaughan’s Tin Pan Alley came on and the ‘demo’ posture dissolved. I didn’t decide to close my eyes. Nobody in the room did. Smiles appeared before words could be formed. After the track ended, Dave turned around to look at me and we all just knew. It was a moment. I bet most of the guys still remember it.

That’s the state we love: time disappears. Our inner critic goes quiet. We stop thinking about the gear and dissecting every aspect of the sound. It’s like being a kid on the playground. Pure enjoyment. This isn’t the absence of detection; it’s what the detection looks like when the body gets there before the inner critic.

It was at that point in the night that we also took a break and reloaded on some snacks, before getting back to the session. But that was it. The SRV moment was gone, and we didn’t get back to that place on that night. It wasn’t the system’s fault, and the other recordings were actually quite good as well.

That’s the fragile part. Tin Pan Alley took us all in, but during the break we drifted into tech talk about bits, sampling rates, tubes, comparisons... Once we were back in evaluation mode, we didn’t recapture that absorbed state later in the same session. Other tracks still sounded good. Other moments still got through. But that high-water mark was behind us.

My practical takeaway after this experience and many others is: don’t mix business with pleasure. If I’m in ‘tweak’ or comparison mode, it’s an evaluation session with my propeller hat and laser level in hand. If I want to disappear into records, I protect a different night for that. Once the scorecard opens, I find it tends to stay open for the rest of the session. I was curious if there was a parallel to this in the recording engineer world, so I asked Christos. He confirmed it:

"I have to step away from a project before I can hear it as a music fan, instead of an engineer."

The analytical tools do their work — we’re being diligent. But this also confirms our need to ‘clock-out’ to fully enjoy the fruit of our labor. Our mental ‘see-saw’ leans on this movement…

The Scenic Route

In taste expertise, researchers found a pattern that might sound familiar. Beginners respond holistically. Then comes the enthusiast phase: more vocabulary, more analysis, more confidence. But also… a greater tendency to miss the whole while staring directly at it. With deeper expertise, the best tasters don’t become less analytical. They become more flexible. They can analyze when it serves its purpose, but can then return to the holistic gist when it’s time to just experience the thing in front of them (LaTour & Deighton, 2019). These studies were not run on audiophiles, but we’ve been running that experiment on ourselves for years. We know the score.

Most of us have lived some version of this. Early on, you just loved music. You may have had no language yet for image density, soundstaging, timing, timbre, tonal balance, or why one system felt more believable than another. You only knew that some nights a record pulled you in and some nights it didn’t. Then the affliction hobby enters your bloodstream. You learn the audiophile jargon. You start hearing deeper into recordings. You also start isolating traits, comparing components, lurking in forums, chasing A/B impressions, building mental checklists, and discovering that the same sharpened attention that helps you hear more can also box you in. This stage matters. It builds the ear. It teaches discrimination. It is how you learn that a system can be spectacular in pieces and still fail to satisfy in whole. But it also creates a recognizable trap: the food critic at the barbecue, the photographer who can’t look at a landscape photo without reverse-engineering the lens used, the audiophile who cannot unhear the squeak in Bonham’s bass drum pedal, the tube hiss in the noise floor, or the last post they read on a forum an hour earlier.

Then, for many serious listeners, something surprising happens. They circle back toward something that may look like the beginning… but it’s not. It’s now familiar ground that’s been earned. The analytical tools are still there, but are now internalized and deployed strategically. They no longer block the entrance. Immersion never went anywhere — the analytical phase just pushed it aside. Evolved experience lets you use both pathways more strategically (Melcher & Schooler, 1996).

This is where Bob Ludwig’s comments resonated for me: he colloquially mentioned that the ‘left-brain-technical’ and ‘right-brain-creative and musical’ are both real demands, but they are not supposed to run full-tilt at the same time. Saturday morning can be for tweaking or taking a new component for a test ride. Another evening can be for wine, a stack of albums or tracks and no agenda at all. Both are legitimate. Both are part of the craft.

If you’re carrying a hammer, you’re always looking for nails. Analysis is a tool, and sometimes you need to learn when to put the tool away. Analytical mode is the chisel. Immersive mode is where you step back to see what you made. What changed is authority. The tools stopped inserting themselves into every session by default. Rigor isn’t rejected… it’s just parked after serving its purpose. It is the last step in becoming serious enough to enjoy what your work uncovered.

This explains the listener side of the ledger. But it doesn’t explain the night Stevie Ray Vaughan entered the room. Everyone in my listening room had heard SRV’s Tin Pan Alley before… we all know it’s a great recording. On a different system it might have been merely ‘really good’. That extra effort I put into tweaking my rig gave that recording exactly what it needed to fully breathe. While the system was still delivering an hour later, that special moment was in the rearview mirror.

Same rig. Different part of the session. We didn’t decide to hear differently; we just did.

Some nights I settle in and the room opens; some nights I can’t quite get there — I get anxious, or restless, and I call it a night early. It is not the gear. Sometimes… it’s just me. By ‘me’, I’m not referring to just mood or mindset — it’s the whole package, including what’s in my blood…

Hippies in the 60s confirmed this 50+ years before researchers did... but for the sake of science, Ferreri et al. (2019) showed how body chemistry can affect people in a double-blind controlled study: they gave listeners two different drugs that affected dopamine. One dampened musical pleasure (risperidone), the other amplified it (levodopa). Our body chemistry affects how we listen.

SECTION 5: The Final Polish

Of course, the whole 'Sometimes… it’s just me' is more than body chemistry. I bring other variables to the game. Maybe I’m set to test a new component or tweak, maybe it’s to dig around for some new music, or maybe I’m just there to go for a ‘ride’. My approach for that listening session can be deliberate, but it can also get hijacked by wildcards I am unaware of. Either way, how we frame a listening session affects how we listen, which aligns with Enstroem & Schmaltz (2022).

We always bring something into the room with us before the first note even plays.

Many of us already know this — it’s why we build routines that might look fussy (maybe even neurotic) to those outside the hobby, but perfectly reasonable from our viewpoint. We usually listen for a few sessions and let it soak in before we flip back to our familiar baseline after a while. We usually know not to put too much weight in that first hour… good or bad. The science behind this made perfect sense to me when I dug-in: when you judge a change against a system you’re already intimately familiar with, you’re anchoring — and with that foundation, the comparison stays more stable than bouncing between two unknowns (Mathias et al., 2020). It’s like comparing a test drive to the car I already own, versus comparing two different models back-to-back that aren’t my daily driver.

Similar to a test drive or reading an article, when we listen to music, the peak and the ending are sorted and weighted differently in our memory (Rozin, Rozin, & Goldberg, 2004). It’s prioritization. It’s efficient. Our brains even start this assembly while we listen — what we hear in the first half is already being updated by what comes after (Mercer & McKeown, 2010). The version of the session we carry forward gets shaped going in, updated as it unfolds, and compressed into a mental ‘zip’ file we reach for later. Each of us takes away a different, very personal sketch of that same scene.

The Chase… and the Hobby

Spec sheets, searching for user reviews, the late-night forum scroll. We’ve all done it. It’s part of the hunt. It’s fun. Then comes waiting for the tracking number, checking it multiple times until it hits the porch, and finally the unboxing and the install preamble for the first power-up and ‘liftoff’. I’m guessing astronauts about to fly into space might be able to relate to how we feel… This is a different excitement than the actual listening. The distinction between 'wanting' (the pursuit) and 'liking' (enjoyment) gives these elements of our hobby a shared ecosystem, without minimizing either side of it (Berridge & Robinson, 2016). In practice, these two elements feed off one another.

The ‘chase’ is the acquisition phase of the listener’s education. The quiet night is where that education gets tested against actual satisfaction. The components or tweaks that don’t survive are just part of the tuition, and the insight we gain from this makes us better listeners. The chase informs the listening. The listening sharpens the next chase. Both are important.

The Anchor

Every new listening session anchors to something I’ve already lived with long enough to know by feel. That’s the reference — not a number on a bench, not someone else’s benchmark, but the one I built in my room over years of calibrating to reference tracks, my personal library, and new music finds. This means ‘my faithful’ was never someone else’s reference... and yours may not be mine.

I believe the concept of ‘fidelity’ has been doing two jobs for us:

Signal fidelity describes how accurately a system preserves the electrical information it is given. Measurement helps us detect a drift, and acts as a benchmark.

Performance fidelity creates both upstream and downstream questions: upstream, when the engineer applies their craft, and downstream — when we tweak our systems.

A familiar example most of us already know: turn the volume up and the sound is re-shaped, it’s not simply a uniform lift across the full spectrum. This reality of how we perceive sound was documented all the way back in the 1930’s when the Fletcher-Munson equal loudness curve was established.

While looking into how the updated loudness curves were built, I learned they were aggregates pulled from young adult listeners with ‘normal’ hearing. The surprise hidden behind these smoothed curves: the reported numbers implied a ~20 dB spread in listener judgments. It would seem that if a microphone is to be the ultimate witness, it requires a ‘jury’ of other microphones…

Sean Olive expanded beyond equal loudness research when he studied listeners’ preference for bass loudness. His numbers showed a 17 dB spread in the level of preferred bass across his test group. At least some of this spread is likely bound to each listener’s unique anatomy: known structures inside the ear proven to shape perception. What struck me about this research is how similar these numbers are to the scatter noted in the equal loudness curves…

Generic ears and averaged plots are approximations… they aren’t individuals.

The auditory measurement world has acknowledged the limitations of its tools, and where real listeners are required to finish the job. In fact, the field’s benchmark textbook by Bech and Zacharov cuts right to the chase: there is no direct measurement for what we actually perceive.

Of course, we all have our own biases, too… But it’s not the whole story. To neatly wrap ‘bias’ around the differences, it would require a stable, fixed target for the ear to be ‘wrong’ about. If there’s one common thread I’ve found in all my years of study and experience on this subject, it’s shown just how slippery that target truly is...

Beyond the measurements, and validation… replicating an experience is harder to pin down. I wondered how this translates to the real-world of mastering, so I went back to my ‘well’ of sound engineering expertise (Christos) and asked what mastering engineers listen to at home, for enjoyment. I suspected they’d want to replicate their studio sound with the same equipment. His response: “Most engineers have a different rig at home because they need something to represent the home environment, not the studio.”

This came as a bit of a surprise to me, so I poked around and found that while some engineers try to take ‘work’ home with them to more closely resemble their studio sound, most use different equipment at home. The bigger surprise to me was many use vacuum-tubed equipment in the chain: Bernie Grundman - McIntosh, Steve Hoffman - Audio Note Ongaku, Bob Ludwig - Manley Steelhead, Dave McNair - VAC Master. The same engineers who are known for crafting some of the best-known recordings ever made, also craft their home systems differently than their studio rigs…

An engineer selects a specific vacuum tube microphone, positions it carefully to get the sound that’s just ‘right’. Decades later, an audiophile fires-up their vacuum tube monoblocs and tweaks the rake angle on their speakers to get locked-in. One of these acts we call ‘craft’. The other?

SECTION 6: A System for One

Dennis arrived and the rest of our audio group started to get settled for the evening. My rig was in that rare place where I’d finally stopped fiddling… Tubes re-biased, subs tweaked… everything buttoned down. Dennis was still early to audiophile gear, but he was as big a music nerd as any of us, and still regularly attended live shows. At some point the subject of ‘best’ bass players came up. I mentioned John Entwistle and Dennis totally lit up… “The Ox!!!”. I verbally nudged him into taking the sweet spot, while I quietly queued-up The Who’s Eminence Front while the group continued to debate the best bass players. I then hit play… let the song run, and the room changed. The background conversation stopped, fingers started tapping, and we were all pulled into the layered Glyn Johns’ mix. The intensity of Entwistle’s bass by mid-track was a force of nature. When the track ended… a brief pause… and then Dennis turned back to me and nodded with a huge grin.

Later, at other group meetups and in other rooms, Dennis would often mention how great that track sounded at my place that evening. His place, somebody else’s, a driveway after the music stopped — he’d circle back to that experience of hearing Eminence Front. There was never a mention of any specific gear.

None of us can recall what tubes were in the amp, what DAC was used, or the last filter settings I spent that summer dialing-in. What’s remembered is how the room went quiet when that familiar track arrived with far more weight than we were ready for, and the stillness afterward…

Most nights don’t become stories. A playlist queues up… and somewhere a few tracks in, I’ve stopped managing the sound. Different nights, different tracks, different people settling into chairs. After a song has filled the room that way… we all know before anyone has to say a word.

Eventually, we queue-up the next track and see where that one takes us...