Jump to content
  • The Computer Audiophile
    The Computer Audiophile

    My Lying Ears

    oop.jpg

    1-Pixel.png

    As a diehard card-carrying audiophile I am interested in all things related to this wonderful hobby. I've published articles based solely on my subjective listening experience and I've published articles detailing only objective measurements and facts about products. I enjoy publishing and reading articles that cover the gamut. I also think it's healthy and interesting to be open to perspectives completely incongruent with our own. With this in mind, I was recently sent a link to the JRiver forum to read a post about one person's perspective and experience as an inquisitive listener. I really liked what I read, in the sense that it's a real world story to which many people can probably relate and it was written in a non-confrontational way. In fact every audiophile I know, golden-eared or not, has at one time or another experienced something very similar to the follow story. I'm not pushing any agenda or endorsing a point of view by publishing this article. I simply think a worthwhile read for all who enjoy this hobby as much as I do.

     

    Here is a a re-written, more complete version of the post, sent to me for publication by the author Michael.[PRBREAK][/PRBREAK]

     

     

    Recently on the Jriver forums (Link) a forum regular was describing his experience at an audio shootout where three bit perfect players were compared. Jriver had not done particularly well in the tests (only receiving 4 out of 38 votes), and there was some discussion of why that might've been the case, given that all three players were (at least notionally) bit-perfect. There were some questions about the test methodology (you can see all the gory details in the linked thread), and some good discussion about how bit-perfect players might or might not conceivably sound different. Ultimately several forum members were of the opinion that the test was basically invalid, while others thought that surely, because so many people heard a difference that there must've been a real difference to be heard.

     

    My own view on this issue is complex. I will confess that I have occasionally heard differences between bit-perfect players. But I don't believe that bit-perfect players actually sound different. That may sound like a paradox, so I'll follow it up with a second one: I don't trust my own ears to correctly detect those kinds of differences in audio. You might well ask “Why not?” Let me offer an embarrassing personal anecdote to explain my point of view about listening tests and the fallibility of the ear:

Several years ago I built a pair of home-made bi-amped speakers. They're each the size of a large washing machine and they took me the better part of a year to build (more than a month of Sundays). Because they were entirely home-made and I was trying to do an active crossover from scratch, even after they were structurally complete, they still required quite a bit of tweaking to get the crossovers dialed in and the EQ set. 

So I started by just dialing in the EQ that seemed to make sense based on the specifications of the drivers, and taking a couple of quick RTAs with pink noise. That sounded alright, and all of my friends (several of whom are musicians and/or “sound guys”) dutifully told me how great they sounded. There was just one hitch: I kept getting headaches whenever listening to the speakers, and the headaches would go away right after I turned them off. So I tried to solve the problem by tweaking some frequencies with EQ. After some tweaks, I'd think I'd made some progress (it sounded better!), and everyone who heard the changes thought the new EQ sounded better.

     

    Eventually, I even started dutifully "blindly" A/Bing new EQ with the old EQ (I'd switch between them during playback without telling my guests what I was switching, which isn't really blind at all), and my guests would invariably swear the new EQ sounded better. And I kept going with this "tuning by ear" method, often reversing previous decisions, backing and forthing and adding more and more convoluted filters. 

The most embarrassing moment (and something of a turning point) was when I was A/Bing a filter, and a friend and I were convinced we were on to something really excellent. After ten minutes of this, we realized that the filter bank as a whole was disabled. I had been toggling the individual filter, but the bank of filters wasn't on, so it wasn't actually even affecting playback at all. And we had been very convinced we heard a difference. And the headaches never went away.

Eventually the headaches (and a growing skepticism) prompted me to stop screwing around and take some real log sweep measurements (at the suggestion of one my more empirically-minded friends). Once I did, I realized that there was apparently a huge (10+ dB) semi-ultrasonic resonant peak at 18.5KHz that I couldn't even actually hear. So I fixed it and verified the fix with measurements. And then my headaches went away. 

This prompted me to take an agonizing look at the rest of the measurements and noticed that my "tuning by ear" which I (and my friends) all felt was clearly superior had turned the frequency response into a staggering sawtooth. So I systematically removed the EQ that was pushing things away from "flat," and kept the EQ that contributed to flatness, and re-verified with measurements. The result sounded so different, and so much more natural that I was embarrassed to have wasted months messing around trying to use my "golden ears" to tune my speakers. And my wife (who had been encouraging, but politely non-committal about my EQ adventure) came home and asked unprompted if I had done something different with the speakers, and said they sounded much better. And she was right; they did. In a few afternoons, I had done more to move things forward than I had in months of paddling around. 


     

    The point of this anecdote is not to try and prove to anyone that my measurement-derived EQ sounded better than my ear-derived EQ or that a flat frequency response will sound best: as it happens, I ultimately preferred a frequency slope that isn't perfectly flat, but I couldn't even get that far by ear. 

The point is that taking actual measurements had allowed me to:


     

    1) Cure my ultrasonic frequency-induced headaches;


    2) Improve the fidelity of my system (in the literal sense of audio fidelity as "faithfulness to the source"); and


    3) Ultimately find the EQ curve that I liked best (which looked nothing like my ear-tuned curve).



     

    My ears (and the inadvertently biased ears of my friends) did not allow me to do any of those things, and in fact led me far astray on issue 2). My ears couldn't even really get me to 3) because I kept reversing myself and getting tangled up in incremental changes. Most damning, my ears were not even reliably capable of detecting no change if I thought there was a change to be heard. 

Once I realized all this, it was still surprisingly hard to admit that I had been fooling myself, and that I was so easily fooled! So I have sympathy for other people who don't want to believe that their own ears may be unreliable, and I understand why folks get mad at any suggestion that their perception may be fallible. I've been accused by many indignant audiophiles of having a tin ear, and if I could only hear what they hear, then I'd be immediately persuaded. But my problem is not that I am unpersuaded: it's that I'm too easily persuaded! I'll concede, of course, that it's possible that I have tin ears and other people's ears are much more reliable than mine, but the literature concerning the placebo effect, expectation bias, and confirmation bias in scientific studies suggests that I'm probably not entirely alone. 

And I've seen the exact same phenomenon played out with other people (often very bright people with very good ears) enough times that I find it embarrassing to watch sighted listening tests of any kind because they are so rarely conducted in a way designed to produce any meaningful information and lead into dark serpentines of false information and conclusions. 



     

    

So to bring things back around: if some bit perfect audio players have devised a way to improve their sound they have presumably done so through careful testing, in which case they should be able to provide measurements (whether distortion measurements on an analog output, digital loopback measurements, measurements of the data stream going to the DAC, or something) that validates that claim. If they claim that their output "sounds better" but does not actually measure better using current standards of measurement, they should be able to at least articulate a hypothetical measurement that would show their superiority. If they claim that the advantage isn't measurable, or that you should "just trust your ears" than they are either fooling themselves or you.

In a well-established field of engineering in which a great deal of research and development has been done, and in which there is a mature, thriving commercial market, one generally does not stumble blindly into mysterious gains in performance. Once upon a time you could discover penicillin by accident, or build an automobile engine at home. But you do not get to the moon, cure cancer, or improve a modern car's fuel efficiency by inexplicable accident. In an era where cheap-o motherboard DACs have better SNR's than the best studio equipment from 30 years ago, you don't improve audio performance by inexplicable accident either. If someone has engineered a "better than bit perfect" player they should be able to prove it, as they likely did their own testing as part of the design process. If they can't rigorously explain why (or haven't measured their own product!), let them at least explain what they have done in a way that is susceptible of proof and repetition. Otherwise what they are selling is not penicillin, it's patent medicine. 

Bottom line: if you and a group of other people hear a difference, there may really be a difference, but there may not be too. Measurements are the easy way to find out if there is really a difference. Once you've actually established that there is a real, measurable difference, only then does it make sense to do a properly conducted listening test to determine if that difference is audible. Otherwise you're just eating random mold to find out if it will help your cough (or headache, as the case may be).

     

    Or you can do what I do for the most part these days: just relax and enjoy the music.

     

     

    - Michael

     

     

     

    1-Pixel.png




    User Feedback

    Recommended Comments



    Empirical evidence is often what drives scientific investigation. To suggest that it has no place is, IMO, akin to sophistry.

    Scientific investigation is what is called for, not unsupported opinions.

    Share this comment


    Link to comment
    Share on other sites

    Miggy this is too complicated for me

    I know... ;)

     

    Could you explain this at my level - the cromagnon subhuman level ?

    Simply out you should be able to measure it in the signal. That's it. If your input USB data has lots of jitter, the receiver chip will have to do a lot of work to fix it. The more work it needs to do the more current it draws from the power bus. Since power busses are not zero impedance this shows up elsewhere. As for single tone vs square, single tone in one frequency. Square is all frequencies - it's possible that higher frequencies get affected more hence the edges of the square would be deformed.

     

    Makes sense?

    Share this comment


    Link to comment
    Share on other sites

    In a sense, there are. They're called confirmation bias.

    Such thing exists. There's also real effects. It is also a bias to decide there can be no effect without testing for it.

     

    We could also induce whatever differences are found by other means and let listeners determine which ones are audible. Then it would be possible to address those problems directly rather than making random stabs in the dark.

    Consider the jitter example. I would say it's not the first thing one would've looked for and yet it turned out to be an important factor in sound quality - a reproducible important factor.

     

    So my take is: if it consistently changes the sound then it must be measurable - what to measure might not be obvious but once isolated one should be able to add/remove the effect and get reproducible results. This is a clearly defined scientific method.

     

    Now ultimately it always boils down to perception in a way since it is listening by which we decide what matters and what doesn't. That doesn't make it any less valid.

     

    I liked the original article bc it said that just perception can lead you astray, but science (ie starting from an objectively flat response) plus a bit of tuning would render much better results than a random walk.

    Share this comment


    Link to comment
    Share on other sites

    +1 Jud your subjective religion is just plain wrong and does nothing but a disservice to progress in the field of audio. All your fancy wordsmithing will never change the facts.

     

    Way to completely miss the point, Sal.

     

    - I agree with Miguel (if you've been paying attention, you may have seen that).

     

    - When you posted, what did I do? Say subjective impressions were king? As John Wayne once said, "Not hardly."

     

    - What I did was ask *you* for some scientific evidence. You say it's not scientifically valid in the absence of an A/B/X test. Fine, as long as A/B/X tests themselves are a scientifically valid way of testing whether people can hear differences between two audio samples. So where is your scientific evidence in support of the validity of A/B/X testing?

     

    - In other words, I'm being *more* demanding of objective evidence than you are. I need objective evidence the test is valid, while you apparently are willing to blindly - religiously? - accept its validity without such evidence.

    Share this comment


    Link to comment
    Share on other sites

    Way to completely miss the point, Sal.

     

    - I agree with Miguel (if you've been paying attention, you may have seen that).

     

    - When you posted, what did I do? Say subjective impressions were king? As John Wayne once said, "Not hardly."

     

    - What I did was ask *you* for some scientific evidence. You say it's not scientifically valid in the absence of an A/B/X test. Fine, as long as A/B/X tests themselves are a scientifically valid way of testing whether people can hear differences between two audio samples. So where is your scientific evidence in support of the validity of A/B/X testing?

     

    - In other words, I'm being *more* demanding of objective evidence than you are. I need objective evidence the test is valid, while you apparently are willing to blindly - religiously? - accept its validity without such evidence.

     

    ROTFLMAO. I got to hand it to you, you're very talented at writing this BS up.

    Share this comment


    Link to comment
    Share on other sites

    If it can't be measured, or at least proven with a properly conducted A-B-X listening test, it will always remain strictly an opinion. An opinion which has no place in any scientific investigation..

    I think it depends on whether it is possible to spot a difference in A-B-X testing.

     

    Let me refer to Chris's original article and the headache listening to his first iteration produced. I bet you a quick A-B-X test would not have spotted that he had a frequency peak up there. In fact maybe a quick test would have make him prefer the case with a peak because it sounds more "airy" in comparison. Clearly the opposite result.

     

    The concept of "blind testing" is something I advocate, but clearly there are more than one way to do that. Like running your system with A for a week, then B for another week, then X for the next week.

     

    I personally would say that I cannot spot specifics in a quick switch situation (unless it's blatantly obvious of course). In fact, most of the time I could not tell you why I prefer one setup over another, I generally realize I enjoy it better - that's it. And it takes me some time to realize this. In some cases after very many back and forths I will be able to focus on what it is that makes the difference, but that's rare.

    Share this comment


    Link to comment
    Share on other sites

    Why does our popular REGEN keep getting dragged into every subjectivist/objectivist argument? What it does has been both measured and heard, so why the controversy? Go pick on magic bricks or something. ;)

     

    (Don't worry, I'm not that naive. I'm just too tired tonight to weigh in with more substantive thoughts on the issue.)

    Share this comment


    Link to comment
    Share on other sites

    I think it depends on whether it is possible to spot a difference in A-B-X testing.

     

    Let me refer to Chris's original article and the headache listening to his first iteration produced. I bet you a quick A-B-X test would not have spotted that he had a frequency peak up there. In fact maybe a quick test would have make him prefer the case with a peak because it sounds more "airy" in comparison. Clearly the opposite result.

     

    The concept of "blind testing" is something I advocate, but clearly there are more than one way to do that. Like running your system with A for a week, then B for another week, then X for the next week.

     

    I personally would say that I cannot spot specifics in a quick switch situation (unless it's blatantly obvious of course). In fact, most of the time I could not tell you why I prefer one setup over another, I generally realize I enjoy it better - that's it. And it takes me some time to realize this. In some cases after very many back and forths I will be able to focus on what it is that makes the difference, but that's rare.

     

    In Chris's case if he had done measurements from the very beginning there never would have been the mystery of the headache. That was exactly the point Chris was trying to teach us. Proper procedure would have been to design the speaker to first measure correctly and then tweak by ear to his satisfaction. It was believing he could depend solely on his ears and not following a correct scientific design procedure that sent him down a rabbit hole. ;-)

    Share this comment


    Link to comment
    Share on other sites

    Why does our popular REGEN keep getting dragged into every subjectivist/objectivist argument? What it does has been both measured and heard, so why the controversy?

     

    Nobody has proved that the measured differences are the cause of the perceived sonic improvements or even suggested a solid mechanism by which they could be.

    Share this comment


    Link to comment
    Share on other sites

    In Chris's case if he had done measurements from the very beginning there never would have been the mystery of the headache. That was exactly the point Chris was trying to teach us. Proper procedure would have been to design the speaker to first measure correctly and then tweak by ear to his satisfaction. It was believing he could depend solely on his ears and not following a correct scientific design procedure that sent him down a rabbit hole. ;-)

    The point I am trying to make is the considering 1 minute A, 1 minute B, 1 minute X as the end-all be-all scientific method is misinformed and wrong. I think this is Jud's point as well.

    Share this comment


    Link to comment
    Share on other sites

    Nobody has proved that the measured differences are the cause of the perceived sonic improvements

     

    True.

     

    or even suggested a solid mechanism by which they could be.

     

    False.

    Share this comment


    Link to comment
    Share on other sites

    ROTFLMAO. I got to hand it to you, you're very talented at writing this BS up.

     

    Then just read what Miguel is writing, since it's the same thing as "this BS," and you seem to be a lot happier reading it from him. :)

    Share this comment


    Link to comment
    Share on other sites

     

    The concept of "blind testing" is something I advocate, but clearly there are more than one way to do that. Like running your system with A for a week, then B for another week, then X for the next week.

     

     

    I like that you put "blind testing" in quotes, because the object is not to be unaware of what one is listening to, but the removal of bias. Blinding is of course one effective way to do that.

     

    A couple of ways to try to eliminate bias other than A/B/X:

     

    - Let people listen sighted to something like speaker voicing or player software (where there are no visual cues), calling the choices something innocuous like "A" and "B." Do this with multiple test subjects. Have them fill out answers to questions about how the voicings sound, differences between them (if any), and preferences (if any). Don't allow the subjects to communicate with each other before filling out the answers. See if there is a consensus, greater than would be expected as a result of chance, about not only preferences but particular characteristics of sound.

     

    - This is a blinded test: Use well recorded monaural music to feed one channel with choice A and the other with choice B. (To try to eliminate room effects, trials could be randomized between right and left channels, and/or headphones might be used.) This eliminates audio memory. Because with monaural music the "stereo effect" (sense that there is one source between the speakers rather than sound coming from each of them) depends on matching the two channels, this is extremely good at identifying differences.

     

    As you say, Miguel, identifying preferences doesn't necessarily establish which of two choices is better from an accuracy viewpoint. Once the fact of an aural difference is established, we can try to identify the cause of that difference in the signal and attempt to measure it to determine which of the different choices is being more faithful to the original.

    Share this comment


    Link to comment
    Share on other sites

    As you say, Miguel, identifying preferences doesn't necessarily establish which of two choices is better from an accuracy viewpoint. Once the fact of an aural difference is established, we can try to identify the cause of that difference in the signal and attempt to measure it to determine which of the different choices is being more faithful to the original.

    Yes. I would steer clear from the term "accuracy" though. The human ear and what "sounds live" is not necessarily flat frequency response. But any difference that one can repeatably hear should be measurable.

    Share this comment


    Link to comment
    Share on other sites

     

    or even suggested a solid mechanism by which they could be.

    False.

    Well, I see that the mechanism JS suggested is argued by John Westlake not to exist

     

    - the crux of his argument is that there is no ramping up mechanism in the USB PHY that is needed to deal with signal integrity issues & therefore no extra current draw, no extra noise generated from the PHY. It's difficult to find detailed information on the PHY at this level of operation & what is the most prevalent method in the field at the moment, in order to evaluate which theory is correct.

     

     

    JW claims that most USB PHY's now operate using an oversampling, clockless data recovery scheme which "When you oversample the input Data, SI Rise & fall times absolutely has no effect on internal circuit conditions, you just need the "Eye Opening" wide enough to recover the data, nothing internally has to "work harder" under poor SI conditions. With USB 2.0 there is no dynamic element unlike USB 3.0 which used a quite intelligent "active" receiver compensation circuits (like HDMI)."

     

    Can anyone throw some light on this?

    Share this comment


    Link to comment
    Share on other sites

    I like that you put "blind testing" in quotes, because the object is not to be unaware of what one is listening to, but the removal of bias. Blinding is of course one effective way to do that.

     

    A couple of ways to try to eliminate bias other than A/B/X:

    .

     

    There is no need to re-invent the wheel, unless of course your trying to influence the outcome

    Double-blind tests are the gold standard in every field of science!

    Share this comment


    Link to comment
    Share on other sites

    Nobody has proved that the measured differences are the cause of the perceived sonic improvements or even suggested a solid mechanism by which they could be.

     

    For many years people said the same thing about jitter--but now everyone knows better. :)

    Share this comment


    Link to comment
    Share on other sites

    For many years people said the same thing about jitter--but now everyone knows better. :)

     

    The effect of jitter is readily measured at the analogue output. Maybe the effects of the REGEN are too, but so far nobody has been able to show it.

    Share this comment


    Link to comment
    Share on other sites

    For many years people said the same thing about jitter--but now everyone knows better. :)

     

    Actually people speculated until someone put together a properly setup test to eliminate bias and use the tried and true scientific method showed there was an audible difference.

     

    Blinded testing is just a measurement tool like a scope, protocol analyzer, microphone, what have you.

     

    I've never seen an audiophile get all lathered up over a mic and pink noise or frequency sweeps.

    Share this comment


    Link to comment
    Share on other sites

    For many years people said the same thing about jitter--but now everyone knows better. :)

     

    What finally happened though is someone setup scientifically rigorous testing and showed that jitter was measurable and blinded listeners could pick it out.

     

    Blinded testing is a measurement tool just like any other. One should get no more bothered about removing bias than using a measurement mic.

    Share this comment


    Link to comment
    Share on other sites

    What finally happened though is someone setup scientifically rigorous testing and showed that jitter was measurable and blinded listeners could pick it out.

     

    I would be very interested in reading about blinded testing of jitter, if you have any references at all you can give, or even a recollection of where and when you might have read about it.

     

    Blinded testing is a measurement tool just like any other. One should get no more bothered about removing bias than using a measurement mic.

     

    Sure. On the flip side, anything that would make the mic less effective would be a concern, and similarly for anything that would affect the ability of the human listener to discriminate.

    Share this comment


    Link to comment
    Share on other sites

    Double-blind tests are the gold standard in every field of science!

     

    That's actually not correct.

     

    Forum member wgscott (an actual scientist in real life) tried to make that point, among others, over at the Hydrogen Audio forums and was pretty well hounded out of them as a result. (I can't remember whether he was banned, or self-banned after talking-tos from HA admins.) And if I were to peg wgscott on an objectivist - subjectivist scale of members of this forum, he'd be well over on the objectivist side, so it isn't that he was trying to sell them some subjectivist claptrap. It's just that he wanted to mention what the actual role of double-blind tests in science is, and they felt he was questioning something that should be unquestionable.

     

    A/B/X is not the only way to do double-blinded tests, and it may be far from the best way, depending on what you are testing for.

    Share this comment


    Link to comment
    Share on other sites

    That's actually not correct.

     

    Forum member wgscott (an actual scientist in real life) tried to make that point, among others, over at the Hydrogen Audio forums and was pretty well hounded out of them as a result. (I can't remember whether he was banned, or self-banned after talking-tos from HA admins.) And if I were to peg wgscott on an objectivist - subjectivist scale of members of this forum, he'd be well over on the objectivist side, so it isn't that he was trying to sell them some subjectivist claptrap. It's just that he wanted to mention what the actual role of double-blind tests in science is, and they felt he was questioning something that should be unquestionable.

     

    A/B/X is not the only way to do double-blinded tests, and it may be far from the best way, depending on what you are testing for.

     

    Correct & citing DBTs as the gold standard is similar to people using the phrase "some people like distortion"

     

    In the field of cognitive science auditory testing has moved onto more robust, accurate methods such as functional MRI, EEG, & MEG with very interesting results.

    Share this comment


    Link to comment
    Share on other sites




    Guest
    This is now closed for further comments




×
×
  • Create New...