Dear JMP Community,
Hoping to hear some thoughts regarding a DOE design targeted at examining standard deviations of a test method.
My colleagues have a test that they perform on a product we make and are wanting to see if they can change parts of the test methodology in order to reduce standard deviations. I'm not aware of a specific DOE platform in JMP that is designed to do this, so we ended up using the custom platform.
After some discussions, we settled on four factors that could be changed: X1 -- continuous factor (-1,1), X2 -- continuous factor (-1,1), X3 -- 2-level categorical (L1, L2), and X4 -- 6-level categorical (L1-L6). We were really interested in each of the four main effects and their first-order interactions: X1, X2, X3, X4, X1*X2, X1*X3, X1*X4, X2*X3, X2*X4, and X3*X4 (10 effects in all). We went with 36 as the total number of runs. But, in order to get an estimate of the standard deviation at each setting, we replicated the DOE a total of 4 times -- for a total 144 test. That was about all we could afford.
We can of course run the DOE evaluation on all 144 tests to estimate the mean response (and noise of the experimentation) with the model factors, which we did. But, what we really want to look at is whether or not a treatment setting changed the test method standard deviation. How we did this was to calculate a standard deviation from the replicated runs (now back to 36 data points) and ran those results through the model.
To me, this all seems like a reasonable way to go about performing a DOE where the goal is to evaluate changes in standard deviations and not means. However, if there is a better approach (maybe from experience), or a DOE specifically designed to do this, we'd be glad to learn about it and perhaps adopt it for the next attempts.
Thoughts and feedback much appreciated.
Thanks!,
DS
P.S. In case you're interested, after evaluating all the individual measurements and the standard deviations from the replicates, we found that factor X4 had the biggest impact on the standard deviations. Unfortunately, when we later ran some verification runs -- now just comparing 32 runs (total) across two of the levels in X4, the results did not support the original DOE and we could not see the improvement in standard deviation we were hoping to (on the plus side, it wasn't any worse!).