Hi Statman,
thanks again for your explanations. I do understand the classical DOE perspective: in a three-factor RSM the model is predefined, the focus is on optimization rather than on model hunting, and things like adjusted R-square, RMSE, and practical relevance are important for adequacy. I am aligned with that.
My problem is not the statistics. It is the practical workload. In my field (protein stability work), I often have ~25 responses per dataset—aggregation levels, charge variants, hydrodynamic size, Tm, etc.—grouped by treatments. The DOE setup itself stays the same (three factors), but the actual response behavior changes with each protein. So I cannot simply reuse a Workflow Builder template; each protein dataset has to be evaluated individually, even though the quadratic model structure is identical.
I also understand contour or surface plots for RSM. These are helpful for intuition. But they are essentially qualitative, because they show the predicted surface in 2D:
y_hat(x1, x2 | x3 = constant)
where
-
y_hat is the predicted response from the model,
-
x1, x2, x3 are the three factors,
-
and “| x3 = constant” just means x3 is held fixed.
This is fine for a single response, but with 20–25 responses it becomes hard to base decisions on contour plots alone.
In practice, I rely more on the Prediction Profiler (and desirability) because it gives me a quantitative way to evaluate and optimize multiple responses at once. For example, for k responses I have predicted values y1_hat, y2_hat, …, yk_hat, and each has its own desirability function d1(y1_hat), d2(y2_hat), …, dk(yk_hat). The overall desirability that I try to optimize is:
D(x) = (d1 * d2 * … * dk)^(1/k)
with x = (x1, x2, x3).
This allows me to numerically optimize the factor settings for multiple stability readouts at once. That’s something contour/error surfaces can’t easily do when the number of responses is high.
So the core issue is not methodology. It’s the time spent clicking. To get from raw data to profiler and desirability-based decisions, I currently need more than an hour per dataset just to run the same quadratic model for each response. Across several datasets, this is four or five hours of mostly repetitive clicking—even though the model definition doesn’t change.
Workflow Builder doesn’t solve this for me, because even if the factors stay identical, each protein dataset yields different responses and I still need to review model adequacy individually.
That’s why I’m looking for a more efficient way—ideally via JSL—to:
-
loop over a list of response columns,
-
apply the same quadratic RSM model in Fit Model and optimize by minimizing for example p or maximizing R2 adjusted,
-
and automatically generate profilers (and desirability)
without having to press “Exclude” again and again.
So my question is simply: is there a way in JMP Pro to batch this process? If you know a JSL pattern or example script that fits this scenario, it would help me a lot. I want to stay aligned with DOE best practice—I just want to reduce manual clicking in the process.