Headphone reviews need a reset

We recently launched our new headphone scoring system and ranking lists, and judging by the reactions, a lot of you were confused. Some headphones you’ve loved for years aren’t ranking as highly as you expected. Some products that routinely receive 8/10 or 4 out of 5 star ratings elsewhere are landing closer to the middle of the pack on our list. So what’s going on?

It’s intentional. And it’s time to explain why.

The Problem With Most Rating Systems

Here’s a familiar scenario. You’re shopping for a new headphone or IEM. You already know the usual forum chaos isn’t going to help much, so you turn to review sites, which inevitably have a scoring system. The product you’re eyeing gets an 8/10, or maybe even a 9/10, four out of five stars, five out of five stars, etc.

That must mean it’s good, right? So you buy it. Eventually, you compare it to something better, or mess around with EQ, and suddenly the headphone that was getting 8 or 9 out of 10 reveals itself to not actually be all that impressive.

This doesn’t happen every time. But it happens often enough. And the missing piece is that the rating scales people happen upon are meaningfully compressed—only using a small fraction of the possible scale of judgment.

If almost everything scores between 7 and 9 out of 10, then the rating isn’t meaningfully differentiating products—it's a 3 point scale, not a 10 point one. This creates false confidence and reinforces confirmation bias. You want the thing you’re excited about to be good, and the rating validates that feeling, and you purchase based on that validation.

Even well-known review sites like SoundGuys and RTINGS—which still absolutely provide useful data—operate within a limited depiction of performance when it comes to their ratings. They don’t always consider the full theoretical range of what headphone sound quality could be, which means the scale homogenizes and compresses. Products that perhaps should fail, often instead get a "soft pass". That’s not good for brands who would benefit from reviewers telling them how to improve their product, but more importantly, its just not good for consumers.

“But Sound Is Subjective…”

Yes, there’s a subjective component to audio, but humans have heads and ears, and that is a fact. There are simply acoustic realities that matter, and a headphone that doesn’t remotely resemble a reasonable ear transfer function isn’t always going to be justifiable as "a different flavor" that will work well for some tastes.

This isn’t about taste. It’s about performance ceilings and performance floors. Right now, the industry operates as though we’ve more or less reached the mountaintop. But we haven’t. But the truth? We're not even close.

You can spend thousands on something like the HiFiMAN Susvara, the RAAL Immanis, or even the Sennheiser HE-1. We’ve all heard them, and none of us feel they represent “as good as it gets.” They all have meaningful compromises.

The Hard Truth: It’s About Compromises

If you take away one thing from this article, let it be this: Buying headphones today is less about achieving perfection, and more about deciding which compromises you’re willing to live with. The HiFiMAN Susvara has treble I personally love—but lean bass and mids that could be stronger. The Audeze LCD-4 delivers incredible bass texture—but has dark upper midrange and feels like strapping bricks to your head. The Sennheiser HD600 has gorgeous vocal timbre—but rolls off in the bass.

Every headphone that I love has trade-offs. I use EQ to mitigate many of the trade-offs—but the point is, we shouldn’t have to.

The sound quality you’re getting right now—even from the best gear in the world—could still be better. The fact that it isn’t simply reflects that the product has compromises, and we want to better reflect those compromises in our ratings.

We asked: Is it fair to score products only relative to what currently exists? Or should we score them relative to what’s possible?

We chose the latter. 

That means the best headphone in the world doesn’t automatically get a 10/10. If it has weaknesses—and they all do—it loses points for those weaknesses. Yes, this results in lower overall scores than you’re used to seeing. That’s the point.

How Our System Works

We have our Sound score, which breaks down sound quality into four categories: Bass, Mids, Treble, and X-Factor. Each category is weighted based on reviewer priorities. For me, treble performance is a make-or-break factor, so it carries significant weight. Another reviewer might emphasize midrange or X-Factor more heavily.

These ratings are subjective listening impressions—not measurement-derived scores. Headphone performance varies from head to head, and no single graph can capture or communicate the totality of our experience.

“X-Factor” can be thought of as technical performance—or, as I sometimes frame it, a bullshit-tolerance coefficient. It’s the intangible quality that makes a headphone better (or worse) than the sum of its parts. 

On top of the Sound score, we have an Overall score that includes Comfort and Value.

What This Means in Practice

Take the HiFiMAN Susvara again. It scores extremely well in treble (8/10) for me. Slightly less in mids (6/10), and slightly less again in bass (5.5/10). That doesn’t mean it’s bad. I love it. But it’s not perfect—and it shouldn’t be treated as if it is, which is why it gets a 7.1/10 for its Sound score.

Or look at the Audio-Technica ATH-R70x. It has some of the best mids I’ve heard (8/10). But the treble lacks precision and detail compared to higher-performing sets (4/10). So its overall ranking reflects that imbalance.

Even strong products land in the 6–8 range because we’re evaluating them against the full performance ceiling—not just against their immediate peers. Meanwhile, truly bad products will not quietly drift by with a 5/10 safety net. If something is bad, it will be treated as such.

A Higher Standard

If you’re a brand and you’re frustrated that your product isn’t getting a 9/10 anymore, understand this: you absolutely can earn that score. It just requires fewer compromises.

Our goal isn’t to tear products down. It’s to raise expectations.

The current media landscape often signals, “This is as good as it gets.” It isn’t.

There’s still enormous room for growth in headphones and IEMs. And if we collectively hold the industry to a higher standard, that growth becomes more likely.

This ranking system reflects that belief. If it feels stricter, that’s because it is. 


This is a companion discussion topic for the original entry at https://headphones.com/blogs/features/headphone-reviews-need-a-reset
6 Likes

Overall, mostly agree. Having said that:

  1. Should the rankings have two different scores? One would be “as is”, and the other score is with EQ applied. I can think of at least two closed back headphones that once EQ is applied, would/should rank higher than a lot of the well regarded open back offerings.

  2. Too much negative bias against closed back headphones in general. Would like to see more effort applied to ranking closed back headphones by applying more effort to EQ them.

  3. Need to consider the hardware driving the headphones more. Some headphones scale up sound better when using higher end electronics.

1 Like

Okay, so where is the reset? I looked on headphones.com for reviews of the Koss PortaPro, and just saw the 2022 review.

How about a table of old vs new with some standards to compare? The PortaPros, HD-650 or 600, maybe a Beyer and an Audeze that are common?

Or play with your tier list showing old and new ratings.

  1. Should the rankings have two different scores? IMO you can just imagine a 9 or a 10 next to every headphone, boom there’s your Post-EQ column :smiley:

  2. Too much negative bias against closed back headphones in general. Did you watch our EQ Tier List video? We’re not interested in ranking based on post-EQ performance (but also, closed back headphones are poor candidates for EQ due to interpersonal variation, despite me personally having a soft spot for them).

  3. Need to consider the hardware driving the headphones more. Some headphones scale up sound better when using higher end electronics. That’s your opinion, I’ve literally never experienced a headphone being “driven so well” that it would shift any of the scores I’ve given to any existing headphones by even a fraction of a point.

2 Likes

Afaik this is the first time we’re doing scores so I don’t think you’re going to find any “old” scores to compare to.

  1. Well, not sure every headphone would necessarily improve. The higher end Audio Technica wooden headphones (EX:AWKT) recommendations are to NOT use EQ, due to it’s internal design.
  2. Whilst most closed back headphones are poor candidates, there are exceptions. The MDR-Z!R sounds amazing with the EQ settings I found for them. I posted them on the site.
  3. I have to respectfully disagree with this. My experience is that some headphones have indeed scaled up and sound better with higher end electronics, especially if using flagship headphones. No way a Oppo HP-1 is going to sound the same as a Topping D900/A900, as an example.
1 Like

At the end of the day, it is a function of trust. Over time I have come to trust the reviewers @ headphones.com. Period.

And it is refreshing to see them reinvent their review approach. They are definitely NOT fools, and I do not think they have any undisclosed financial incentives for doing so, except of course to improve their reviews.

Only thing I would ask, is that they review the benefit of adding an impulse response measurement, for all headphones and IEMs. In my experience, this tends to correlate the most with how “accurate” a listening devices sounds. My experience of this comes from speaker reviews, especially the ones produced by resolutionmag.com (which is no longer active), but maybe the archives are available at archive.org. I expect the same should apply to head worn listening devices.

unheardlab.com also uses impulse responses in his reviews.

Example from his FT1 Pro review at this page/link

Chaps, please continue doing what you do best.

2 Likes

Can anyone please point me to a good headphone.com or related Youtube video on how to EQ headphones?

Thanks

For starting I found this video from the master of EQ helpful:

Actually, the RTINGS headphone evaluations are pretty informative. Lots of measurements are provided. ASR also provides fairly detailed information that provides insight regarding sonic character. The music genres can change the evaluation scores. People to listen to primarily classical/acoustic music are likely to have a different scoring criteria for headphones than those to listen to rock/electronic music.

Evaluating headphones is rather challenging. Even the scores provided by the evaluation team differ to some extent. There are so many headphone offerings available, it’s difficult to provide a comprehensive list. For example, there are headphones that ASR recommend that are not even listed here.

The problem with ASR is that they’re basically using headphone measurements incorrectly to judge headphones (some of the users there know this). They measure headphones on GRAS rigs with KB5000 pinna shown against the Harman target, but what they don’t realize is that the target was never devised on that pinna, it was devised using the Welti-modified pinna and the results from that research show notable differences in HpTF behavior for those headphones. So they’re effectively saying “it must match this target, because this is the best research”, not realizing the tools they use to get the data aren’t appropriate for that research.

Somewhat amusingly they criticise the use of more advanced measurement rigs because “it’s incompatible with the research”, when the very systems they use are also incompatible with that research.

Beyond that… no matter how many times Dr. Olive says “it’s not meant to be used that way”, they continue to be militant about target adherence.

There are of course all kinds of other problems with that methodology, using a highly smoothed target, using a single line target to begin with, not considering acoustic impedance and load sensitivity, making judgments on target adherence based on the most absurd data normalizations… I could go on. It’s a narrative farm that’s developed its own form of snake oil and called it science.

4 Likes

Shots fired! :joy:

In the video version of this, you mention that some might see your position that headphones could be a lot better than they are as “arrogance”. I’ve been thing about this for a few days and I have to say that, while I agree with what you are doing for the rating/score, it does still strike me as a bit arrogant to say that a no-compromise headphone could be made - even theoretically. For example we can’t just add EQ the headphone cable to make it sound exactly as intended without compromise. And even if we could, it wouldn’t be the right EQ for every listener. So the only 10/10 headphone would have to include its own DAC/amp for DSP and a built-in way to measure a detailed HRTF. And now it’s surely compromising on comfort and/or value.

Oh god! I just described Bose QC Ultra Gen 2! :grimacing:

My conclusion is that if it were up to me I’d continue with your new rating system, but I’d change tier list to consider S tier to be the best available and not a theoretical “near perfect” tier. The main reason is that I want a tier system to rank the available headphones against each other. And also the B tier is too crowded. But that’s just my 2 cents.

1 Like

I can understand why it may seem like that’s what we’re doing by rating things lower, especially since obviously we’re big EQ proponents, but I actually think that’s not how we feel about it - certainly speaking for myself, I’d still draw a distinction.

For the headphones that currently exist, it’s true that I’d need to EQ them to get 10/10 sound. But I think we’re not yet at the point where the last bastion of sound quality possible in a product is EQ, or a CustomTune type technology for the entirety of the ear response. I can imagine a passively tuned headphone that achieves say… 9/10 for me, and I know this because there are headphones that for one aspect or another perform rather well. I think I do have 9s on the board for certain aspects of several headphones.

I was having exactly this conversation with a headphone product manager of a very large company (someone who was also rather influential in the science side of stuff in the past. I won’t say who). I was telling him how I end up EQing all of my headphones and he said “yes, people like you and me will always do this”, implying that ‘normal’ people will always have to endure compromises. There’s an inherent arrogance to this in that “the masses don’t get to hear what I get to hear, they’re having worse experiences, so only I KNOW what sound can be”, and so on. When considering the scale this person was working with, I do have to agree it’s not realistic, and for that, our ranking list standards are likely too high.

But our space also caters to people looking for experiences without compromises, where there are brands like HEDD audio, Meze, DCA, Sennheiser etc., with extremely knowledgeable and skilled engineers who are trying to achieve that. When I talk to them about this stuff, they know exactly what we’re on about, because they’re trying to deliver it too. I believe that for their customers, they can get a lot closer than they have so far. And it’s not just a matter of design choices, or knowing what good is, it’s a question of production, being able to bring that vision to market in a way where the outcome matches what you wanted to achieve.

It’s not easy, but I know the people working on this stuff also typically feel more can be achieved, even without EQ. We’re not saying anything they don’t already know, and I’ve had several conversations with manufacturers about this already where they say yeah, we can do it. So needless to say, there is room to improve, and I’m optimistic about it.

3 Likes

Appreciate the feedback. Still learning the unique aspects of headphone science.

I do find it hard to square some of the grades assigned compared to my in-situ preferences. I’ve owned several Focal headphones, and wound up selling all of them. I kept reading about how it should sound, but just couldn’t warm up to them. I found the Sennheiser “house sound” to be far more enjoyable. Also found the ADX5000 to be more to my liking than the Focal Clear/Clear MG.

One thing I’ve learned over the years is that a review only becomes meaningful when you understand how the reviewer listens — their chain, their references, their priorities, and the language they use to describe what they hear. Without that context, even well‑intentioned reviews can feel disconnected from a reader’s actual experience.

For me, the most helpful reviews — written or video — are the ones that combine clear methodology with perceptual honesty. Not ‘this is the best,’ but ‘this is what I heard, here’s how I listened, and here’s why someone else might hear it differently.’

2 Likes

Not to mention that at least some reviews listening sections are performed at excessive volume.

I, at least hear pear to oval tones, with a hint of walnut. On the pinnae, notes of plum with an undertone of turkey broth.

1 Like

II enjoyed reading your lists and seeing where some of my favorites landed. Fun read!

You have far more knowledge about how the sausage is made, but I’m still skeptical a headphone could get 9’s across the board with a 10 in there somewhere - without EQ.