Hello again, @HoaLe99 !
This is a really interesting example! I've just spent the last 2 hours exploring this when I really should be doing other things. But it was so much fun. Thank you for sharing!
First of all, about the threshold: I wasn't recommending that you change the p-value threshold in the analysis. I think you can do that, but it wouldn't change anything. You would still get exactly the same analysis and p-values. I was really just saying that we declare "significance" when a p-value is below the threshold of 0.05. And that threshold is only a convention or a "rule of thumb". We could easily use a different threshold and that would change what we say is "significant."
Anyway, it is largely irrelevant in this case because the p-value for lack of fit is a long way below 0.05. It really does look like you have fairly serious lack of fit, which you can see from some of the plots:
Notice how the points are not just randomly distributed around the line in the actual vs predicted. There are groupings above and below the line at different points.
This suggests that the model is inadequate in some way, so it might require higher order terms. You only considered first and second order terms (the RSM model). I looked at adding 3rd order terms.
This is complicated because there are more 3rd order terms than you can fit. I actually used some of the tools in JMP Pro to help with this, including Best Subset and SVEM Lasso in Gen Reg in JMP Pro 17. However, you can use Stepwise in standard JMP. I locked in all the second order terms and then did forward stepwise to see which 3rd order terms were most important. (There is a script for Stepwise in the attached version of the table)
It also seemed to me that your response should not be less than zero, and possibly not greater than 1, so I added a Logit transform to the response for the "best" model (script in attached table). Without this transform you can get response predictions below 0, which I suspect is not possible here.
I found that the best model was the RSM with the cubic term for X2 (X2*X2*X2) and the interaction of the X3 with the quadratic of X1 (X1*X1*X3). This is a complicated effect. It means that the curvilinear nature of X1 is affected by X3. The curve of X1 is more pronounced when X3 is higher. Maybe that makes sense scientifically? It is unusual to fit these in models of experiments, but my feeling is that they are probably quite important effects in many cases.
The model still has a significant lack of fit, but it is much less (p = 0.009 versus 0.0004). And the residuals all look better to me. To improve the model further you would probably need to augment with more runs to test the 3rd order terms. I am not sure that this would be worthwhile but that depends on what you need the model for.
This is a really interesting example, so please let me know if this will ever be published or if it would be okay to use this in anonymised form.
I hope this was helpful for you. I suspect the model that you already had would have been useful for your objectives. The improved model looks better statistically but the difference might not be that important practically speaking.
Phil