OT, but here's an interesting "double-blind" test that probably wouldn't pass the Hydrogenaudio TOS test either:
http://www.npr.org/blogs/deceptivecadence/2012/01/02/144482863/double-blind-violin-test-can-you-pick-the-strad
Was your comment a veiled shot at hydrogenaudio, did you not read the article, are you questioning the test's adherence to double blind testing, or are you unfamiliar with hydrogenaudio's ToSs? The violin test was specifically mentioned as being "double blind", which obviously meets HA's ToS for claims if the test actually was double blind, and I've not seen anyone anywhere (other than you) question the validity of the test's methodology, analyses, or conclusions.
But maybe you've caught something that peer review of the published study overlooked - please provide your evidence that the test does not meet the standards for a valid double blind test or other deficiencies in the study.
This NPR article made me think of recent discussions concerning DBX testing on Interact, and gives an instructive example of some of the difficulties inherent in setting up a good DBX test.
First, it's important to note that there is a DBX test of two audio files embedded in the article itself, and that's not the one I'm going to complain about! It's a cool test, however: two posted excerpts feature the same musician playing the opening to the Tchaikovsky violin concerto, one on a violin made in 1980 and the other on a Stradivarius violin (made probably around the year 1720). There are no results listed for this test, but I personally hear a clear difference between the two instruments on the recordings. In fact, one sounds hands down better to me. This seems like classic DBX testing setup, and it's fun...take the test and see if you can pick the Strad!
However, the NPR article's description of the actual playing test conducted by researchers and players doesn't seem like a fair at all. First, there are six instruments being evaluated at once, which seems absurdly large to me. It's very hard to evaluate and remember sound differences between two samples, let alone six!
It's hard to tell what constitutes a successful answer from reading the article, but it sounds like the musicians had to get all three Italian instruments correct, otherwise they "couldn't tell the difference." Statistically, would anybody ordinarily get three answers correct randomly out of a test group of 17? Perhaps the fact that 3 people did identify the instruments points toward a substantive difference in the instruments, not away from one!Also, there are lots of variables that aren't mentioned in this article. First, setup of these instruments with strings, post, and bridge is a huge variable. I would assume care was taken to minimize this factor, but it is impossible to eliminate. Second, many of the great, older string instruments sound better in a large hall, and under the ear sound undistinguished. I've experienced this effect firsthand, walking away from a Stradivarius cello being played in a concert hall and hearing the sound get louder, richer, and warmer as I walk into the acoustic of the hall.
The consensus among musicians that have played, owned, or worked with Stradivarius instruments is that they offer unique, distinguishing tonal qualities that can't be found elsewhere. I would demand a better test and more evidence before that gets turned around, and the test described in this article doesn't convince me at all.
What does all of this have to do with Audio DBX testing? Ha ha! Well, everything...and nothing.
I'm not really that familiar with Hydrogen Audio boards, other than my reading of the responses to the Absolute Sound computer audio article posted by JimH on that board and studying their terms of service. It doesn't sound like this "Audio" test in the NPR article would fly (that's assuming the subject at hand, fine Italian musical instruments, were of interest to the members in the first place!)
Basically I think constructing a good DBX test with violins is virtually impossible, there are just too many variables at play. Therefore, I'm inclined to believe that DBX is of limited use in making evaluations about music. The conditions necessary to isolate variables are so hard to achieve that they limit options by default.
This is certainly worth considering in the Music vs. Audio vs. Science debate that emerges on forums out of debates like Wav vs FLAC, bit perfect sounds different, etc. Science is science and engineering is engineering, but audio systems aren't simple; isolating variables for a good DBX is very difficult. I've certainly heard differences in digital cable lengths, etc. come out of changes in super resolving systems using big Martin Logans, etc.
Anyway, food for thought!