Peruvian802
Well-Known Member
After pursuing this thread, I think I am finally ready for my probabilities and statistics final. Thanks everybody!
https://www.homebrewtalk.com/showthread.php?t=633396
It's about how what you have drunk prior can affect your perceptions of later beers. Which, of course, is the single biggest issue with the triangle testing as it is presented. It's why "qualifying" people on the basis of a lucky guess doesn't make any sense, and why there may be more differences perceived by a panel if they had done the testing under more controlled conditions.
Thus it depends on whom and what you are interested in. The fact that assessors trained to tell the difference between Beer A and B can tell the difference between beer A and B doesn't tell me much about whether my SO is going to be able to tell the difference. OTOH if that panel was trained (or pruned) to detect diacetyl then I have confidence that Beer C is different from beer D with respect to diacety if they determine it to be so. The problem here seems to be that the investigators have not done all their homework in terms of determining what they want to determine before doing the experiments.Choose assessors in accordance with test objectives. For example, to project results to a general consumer population, assessors with unknown sensitivity might be selected. To increase protection of product quality, assessors with demonstrated acuity should be selected.
But in general, I feel that the meta-analysis in this case has bolstered the contention that fermentation temp has a statistically significant effect on beer.
Because you will pass every single time you do the test. That's why I stopped using it. I have never done a diastatic power calculation and never had a mash that has failed to convert completely. What's the point of doing a test that tells me I'm completely converted when I already know that?
https://en.wikipedia.org/wiki/Meta-analysis
Interesting. Based on the Wiki article, it seems that it is okay to combine. Clearly there wasn't any problem in the searching for studies (as this was covered by them all being run by the same group), and there isn't a publication bias problem because we know that Brulosophy publishes both results and non-results. Given that they used the same methodology in each study, that means that I didn't have to correct for multiple results to reach a homogeneous sample--they were already homogeneous.
However, the aggregation of studies is still limited by the quality of the original experiments. If there were flaws in their methodology (such as not doing AAB, ABA, BAA, ABB, BAB, BBA random presentation of samples, etc), that error is not in any way corrected for by the meta-analysis.
One could also accuse me of bias because I already believe that temperature control is important, but I could not use that bias in any way given my selection of studies. I took all 7 of the Brulosophy studies comparing cold vs warm ferment as written without qualifying them, and the only study I excluded--which was static vs variable temp--actually had a stronger result than studies I used so excluding it didn't strengthen my case.
But in general, I feel that the meta-analysis in this case has bolstered the contention that fermentation temp has a statistically significant effect on beer.
I guess I was thinking of doing it at 20-30 minutes and sparging from there if the conversion is done. that is one way to save a few mins on brew day![]()
Five different yeasts. That's right five. Three different people testing it in different states. 8 experiments in total, 6 unable to show significance as tested. One of them brought to National Homebrew convention and tasted by everybody, including famous people in Homebrew. Tons of anecdotal and qualitative data. Meaningless preference data often times showing preference for the warm fermented anyways. And yet you have added everything together so you can prove something. I think you've done a great job of showing how real information can be skewed in any way anybody wants. You are a smart guy, I dont get what you are holding onto so much that you feel the need to add all these negative results together to reach a positive. Look I didnt care if any of this was true from the get go, so I guess we just had opposite beginnings. It seems way to upsetting to me. If someone told me pasta could be made in cold water, it wouldnt make me all angry either. Beer can be mashed in cold water overnight it turns out. I think its cool, not something to disprove. Brulosophy is cool and interesting to me, i try not to attach any personal value on the findings.
Sometimes I just like to geek out on numbers... I'm an engineer, after all
This one was particularly interesting to me, though, as I've always thought that fermentation temp control was a significant step in improving my beer. I even thought that years ago when I did a double-batch with my former brewing partner, and we fermented one in a temp-controlled fridge but had to put the other one in his spare bathroom un-controlled. The un-controlled batch seemed "hot", i.e. with higher-alcohol flavors.
But for me, something just seemed "off" regarding these experiments. While the individual experiments didn't achieve significance, I was noticing that the error was always in the "correct" direction. From a statistics standpoint, that belied the idea to me that we weren't dealing with blind chance. It suggested that beer was not completely indifferent to fermentation temperatures, but that perhaps the effect was too small to be seen based on the size of the tasting panels.
That's why I looked at the meta-analysis, and based upon what I can see, my tactic was NOT statistically unsound. Meta-analysis is used for just this purpose--to find effects that may be too small to be significant in individual experiments, but taken in the aggregate are meaningful.
FYI there's now a 9th ferm temp experiment, and this one actually hit p=0.002... Interestingly this one was taking WLP300 and fermenting the batches at either 60 or 72, both within range.
So with this, we're now at 91 of 208 (43.75%) correct identification of the odd beer in the triangle, against an expected null hypothesis of 33%. This corresponds to p=0.001. Assuming meta-analysis is a valid technique, these experiments are absolutely demonstrating IMHO that fermentation temperature affects the finished product.
But of course, as you state, that doesn't get to preference. Which is obviously an important factor, and the warm ferment lager was preferred to the cool, which is a surprising finding.
But in general, I feel that the meta-analysis in this case has bolstered the contention that fermentation temp has a statistically significant effect on beer.
Lets look at the combined data and see what it implies:So with this, we're now at 91 of 208 (43.75%) correct identification of the odd beer in the triangle, against an expected null hypothesis of 33%. This corresponds to p=0.001. Assuming meta-analysis is a valid technique, these experiments are absolutely demonstrating IMHO that fermentation temperature affects the finished product.
Since I'm advocating ASTM E1885 as gospel let me quote a bit from ¶7.2
Thus it depends on whom and what you are interested in. The fact that assessors trained to tell the difference between Beer A and B can tell the difference between beer A and B doesn't tell me much about whether my SO is going to be able to tell the difference. OTOH if that panel was trained (or pruned) to detect diacetyl then I have confidence that Beer C is different from beer D with respect to diacety if they determine it to be so. The problem here seems to be that the investigators have not done all their homework in terms of determining what they want to determine before doing the experiments.
Man, I'm not as nerdy into statistics as a few of you guys by a long shot, but I nerded out on this thread with you. Thanks AJ and bwarbiany for the great posts. My inclinations were initially the same as yours B, but I didn't have the stats background to show it. And then as AJ pointed out, it has to actually reach the null hypothesis in order to be considered fully indifferent (if I understood that correctly). So although the experiments didn't *prove* that fermentation temp mattered, they didn't *disprove* the theory either. Rather, they should've raised more questions and more testing. Then the fact that the meta-analysis is actually acceptable practice, and that it, in fact, does *prove* that fermentation temp will make a difference, only furthers my gut-feeling that if the panel sizes were larger, we would see more significant results.
I'm not sure how inclined AJ is to converse with the brulosophy dudes, but just from my random interactions with them online, I have a feeling they'd actually appreciate some of this stuff. I would bet you that they're much closer to the mode of thinking of wanting to actually do things right, and would prefer to have such inclinations as fermentation temp proven correctly. And I'd venture to say that they're completely against those types in this thread who are taking these experiments as brewing gospel.
Lastly, after going through the meta-analysis on the ferment temp experiments, it'd be interesting to do the same with others that are testing the same variables.
Man, I'm not as nerdy into statistics as a few of you guys by a long shot, but I nerded out on this thread with you. Thanks AJ and bwarbiany for the great posts. My inclinations were initially the same as yours B, but I didn't have the stats background to show it. And then as AJ pointed out, it has to actually reach the null hypothesis in order to be considered fully indifferent (if I understood that correctly). So although the experiments didn't *prove* that fermentation temp mattered, they didn't *disprove* the theory either. Rather, they should've raised more questions and more testing. Then the fact that the meta-analysis is actually acceptable practice, and that it, in fact, does *prove* that fermentation temp will make a difference, only furthers my gut-feeling that if the panel sizes were larger, we would see more significant results.
I'm not sure how inclined AJ is to converse with the brulosophy dudes, but just from my random interactions with them online, I have a feeling they'd actually appreciate some of this stuff. I would bet you that they're much closer to the mode of thinking of wanting to actually do things right, and would prefer to have such inclinations as fermentation temp proven correctly. And I'd venture to say that they're completely against those types in this thread who are taking these experiments as brewing gospel.
Lastly, after going through the meta-analysis on the ferment temp experiments, it'd be interesting to do the same with others that are testing the same variables.
The probability that Marshall is reading this thread is 1.
Although I've worked with a scientific modeler who says that there's never a probably of 0 or 1...
Well I certainly wish I'd remembered that post. I'd be $45 richer! As to the credibility of the source it is, by definition, credible as it is a standard. There are typos in it and I think I have found an error in the confidence interval calculation which I did not correct in the spreadsheet because it is a standard (and I'm not, at this point, 100% sure i'm right).Hey AJ this seems to be the same E1885 I linked back in post #25. Guess you answered my question about credible source. Rang a bell cause we both pointed out section 7.2. It's not $45 it's free.
In the one description of an experiment I read all panelists were presented, at best, the permutations of AAB (no BBA). Panelists that couldn't decide were not instructed to guess. These are pretty glaring errors in protocol and suggest that there were others. For example, I doubt that they have individual isolated booths for their panel members. These things would introduce 'noise' yet the single experiment and the pooled experiment results both suggest that the differentiability is about 15% so I think we have to conclude that they did some things right.Oddly my reading of same document pointed at how much the brulosophers were getting right in their effort to use the triangle test while your read is more or less the opposite.
I think you are doubtless right but even the single case experiment I looked at (9 out of 21 correct guesses) supports that conclusion. Though those results only support rejection of the hypothesis that fermentation temperature does not make a difference to the 25% significance level that does not prove the null hypothesis is false. In the long post I showed that we can calculate the probable range of differentiabilities and these data indicate it is between 0 and 0.35 with 90% confidence. Based on that we are not likely to accept the null hypothesis (Pd = 0). I took it one step further and modified the calculation routine to also calculate the most probable value of Pd given the number of correct answers obtained and the panel size. For 9 out of 21 correct this is Pd_max_liklihoood = 14%. That's respectably far from the hull hypothesis Pd = 0.
H1(Pd=0.30): Panelists: 21; 3-ary test; 9 Correct Choices; P(< 9) = 0.11886; 0.00 < Pd < 0.35 with conf. 0.90
Most Likely Pd: 0.140
Lets look at the combined data and see what it implies:
H0(Pd=0): Panelists: 208; 3-ary test; 91 Correct Choices; P(>= 91) = 0.00112
H1(Pd=0.20): Panelists: 208; 3-ary test; 91 Correct Choices; P(< 91) = 0.18079; 0.09 < Pd < 0.22 with conf. 0.90
Most Likely Pd: 0.160
The larger sample size gets us to confidence of 0.0011 that we can toss the null hypothesis. Fiat! It also tightens our 90% confidence band for differentiability (and gets 0 out of it) and gives us a maxiumum liklihood estimate of 16% for the differentiability as opposed to 14%. That's not really very different. This consistency suggests that the various experimnental result are indeed capable of being combined. But what did it buy us? A substantial improvement in support for rejecting the null hypothesis. But we already were pretty sure we should do that. Or, put another way, alpha bigger than we might like based on the assumption that it always needs to be < 0.05 isn't the whole story.
Excellent response. So what did you take from that xbmt? I took a lot. It confirmed what I had been thinking for a while that ale, hef, some yeasts are more reactive then others. I certainly never meant or intended to be pigeonholed into any yeast strain never showing a difference. Just that it didnt always matter in the grand scheme of things.
Imo the assumption is in defense of the purchased equipment. Quickly the assumption seems made that there was a difference, so yay cold was better, i am right, this is how good beer is made. But putting oneself in not caring shoes, one can see that 7 preferred the warm and 8 the cold, 4 had no preference. Furthermore the brewer claims taking a 3rd place medal with a 75/25 mix of the warm and cold. So should I go buy a fridge and controller? If so why? Is anyone really surprised that a hef yeast showed a reaction to temp? Which one is better? The cold, right because that is the common thought and equipment defense. Surely, this has to make sense to someone.
So they noticed a difference, now what? If you like hefs, it says to me that flavor could be manipulated by temp and that warm and cold would both be worth trying. It says to me that 72 would be ok for a hef as preference seems split anyways and he did well with the 3rd. Nothing in this speaks to dogmatism about ferment temp in the grand scheme of things. Both warm and cold will make good beer and the joy of not caring or extra equipment gives the warmer an edge to me.