<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic removing terms from a model following a designed experment in Discussions</title>
    <link>https://community.jmp.com/t5/Discussions/removing-terms-from-a-model-following-a-designed-experment/m-p/703191#M88747</link>
    <description>&lt;P&gt;Ok, suppose I've conducted a full factorial design with four 2-level factors. When fitting the model I've included all 2-way interactions. Some of the factors' main effects and interactions are not significant at 0.05. If I then remove the least significant factors (those at the bottom of the effect summary table), some of the non-significant terms become significant. If I were trying to develop the 'best' model, I would feel comfortable doing this. But if I am trying to determine whether a factor plays a role in the response variable then I'm not sure this is legit. Is there a proper rationale, in a DOE context, for removing insignificant terms and then describing the remaining terms as significant?&lt;/P&gt;</description>
    <pubDate>Mon, 27 Nov 2023 15:10:42 GMT</pubDate>
    <dc:creator>gchesterton</dc:creator>
    <dc:date>2023-11-27T15:10:42Z</dc:date>
    <item>
      <title>removing terms from a model following a designed experment</title>
      <link>https://community.jmp.com/t5/Discussions/removing-terms-from-a-model-following-a-designed-experment/m-p/703191#M88747</link>
      <description>&lt;P&gt;Ok, suppose I've conducted a full factorial design with four 2-level factors. When fitting the model I've included all 2-way interactions. Some of the factors' main effects and interactions are not significant at 0.05. If I then remove the least significant factors (those at the bottom of the effect summary table), some of the non-significant terms become significant. If I were trying to develop the 'best' model, I would feel comfortable doing this. But if I am trying to determine whether a factor plays a role in the response variable then I'm not sure this is legit. Is there a proper rationale, in a DOE context, for removing insignificant terms and then describing the remaining terms as significant?&lt;/P&gt;</description>
      <pubDate>Mon, 27 Nov 2023 15:10:42 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/removing-terms-from-a-model-following-a-designed-experment/m-p/703191#M88747</guid>
      <dc:creator>gchesterton</dc:creator>
      <dc:date>2023-11-27T15:10:42Z</dc:date>
    </item>
    <item>
      <title>Re: removing terms from a model following a designed experment</title>
      <link>https://community.jmp.com/t5/Discussions/removing-terms-from-a-model-following-a-designed-experment/m-p/703243#M88753</link>
      <description>&lt;P&gt;Here are my thoughts in general (specific advice would require a more thorough understanding of the situation):&lt;/P&gt;
&lt;P&gt;First, I will assume you have an un-replicated factorial with 16 treatments. &amp;nbsp;The model includes all possible terms to 4th order.&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;You bring up an important concept with respect to experimentation and the F-test (or any test perhaps). &amp;nbsp;The typical significance test in experimentation is to compare the MS of the model term with the MSe (error). &amp;nbsp;This is the F-ration or F-value. The important questions are: How was the error term estimated? &amp;nbsp;How representative of the true error is the estimate? &amp;nbsp;Are the comparisons that are being made useful and representative?&lt;/P&gt;
&lt;P&gt;If you remove insignificant terms from the model (lack of fit), you are potentially biasing the MSe lower (small SS divided by DF). &amp;nbsp;Hence when you compare the MS of the model term with the smaller MSe you get larger F-values (and smaller p-values).&lt;/P&gt;
&lt;P&gt;A quick read:&lt;/P&gt;
&lt;P&gt;&lt;A href="https://www.additive-net.de/images/software/minitab/downloads/SCIApr2004MSE.pdf" target="_blank" rel="noopener"&gt;https://www.additive-net.de/images/software/minitab/downloads/SCIApr2004MSE.pdf&lt;/A&gt;&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;I recommend first assessing practical significance of the model terms using Pareto charts of effects AND use Daniel's method of evaluating statistical significance for un-replicated experiments (Normal/Half Normal plots) and perhaps augmented with Bayes plots (Box). &amp;nbsp;This will give you both practical significance and statistical significance without bias.&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;Daniel, Cuthbert, "Use of Half-Normal Plots in Interpreting Factorial Two-Level Experiments", &lt;EM&gt;Technometrics&lt;/EM&gt;, Vol. 1, No.4 November 1959&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;Once you have determined the factors/interactions that are active in the experiment, simplify/reduce the model. &amp;nbsp;The purpose of simplifying the model is 2 fold:&lt;/P&gt;
&lt;P&gt;1. to get a more useful model for iteration and prediction&lt;/P&gt;
&lt;P&gt;2. to get residuals to help assess model adequacy and whether any assumptions were violated.&lt;/P&gt;
&lt;P&gt;&lt;STRONG&gt;Note: you do not re-assess statistical significance for the simplified model&lt;/STRONG&gt;&lt;/P&gt;</description>
      <pubDate>Mon, 27 Nov 2023 16:39:08 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/removing-terms-from-a-model-following-a-designed-experment/m-p/703243#M88753</guid>
      <dc:creator>statman</dc:creator>
      <dc:date>2023-11-27T16:39:08Z</dc:date>
    </item>
    <item>
      <title>Re: removing terms from a model following a designed experment</title>
      <link>https://community.jmp.com/t5/Discussions/removing-terms-from-a-model-following-a-designed-experment/m-p/703252#M88754</link>
      <description>&lt;P&gt;Thanks. Your point about biasing the MSe is at the heart of my concern about my conclusions about the significance of the remaining terms ... that I probably have a higher type I error rate than I think I do. So I may be falsely claiming a factor effect.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;I'll add that it was a replicated experiment with four replicates, since we expected poor signal to noise. Does that help guard against false claims of term significance?&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;The factors were binary/categorical and the response was continuous.&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Mon, 27 Nov 2023 16:48:16 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/removing-terms-from-a-model-following-a-designed-experment/m-p/703252#M88754</guid>
      <dc:creator>gchesterton</dc:creator>
      <dc:date>2023-11-27T16:48:16Z</dc:date>
    </item>
    <item>
      <title>Re: removing terms from a model following a designed experment</title>
      <link>https://community.jmp.com/t5/Discussions/removing-terms-from-a-model-following-a-designed-experment/m-p/703286#M88757</link>
      <description>&lt;P&gt;To add some remarks and comments in addition of &lt;a href="https://community.jmp.com/t5/user/viewprofilepage/user-id/4358"&gt;@statman&lt;/a&gt; answer.&lt;/P&gt;
&lt;P&gt;&lt;BR /&gt;Your topic seems very similar to models evaluation and selection. In this process, statistical significance may be one metric, but not the only one (always add domain expertise !). &lt;BR /&gt;Complementary model's estimation and evaluation metrics, like log-likelihood, information criteria (AICc, BIC) or model's metrics (explanative power through R2 and R2 adjusted, predictive power through MSE/RMSE/RASE, ...) offer different perspective and may highlight different models.&lt;BR /&gt;You can then select based on domain expertise and statistical evaluation which one(s) is/are the most appropriate/relevant for your topic, and choose to estimate individual predictions with the different models (to see how/where they differ), and/or to use a combined model to average out the prediction errors.&lt;BR /&gt;&lt;BR /&gt;"All models are wrong, but some are useful" is a quite adequate quote when dealing with multiple possible models and trying to evaluate and select some to answer a question.&lt;BR /&gt;&lt;BR /&gt;Hope this additional answer may help you,&lt;/P&gt;</description>
      <pubDate>Mon, 27 Nov 2023 17:08:35 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/removing-terms-from-a-model-following-a-designed-experment/m-p/703286#M88757</guid>
      <dc:creator>Victor_G</dc:creator>
      <dc:date>2023-11-27T17:08:35Z</dc:date>
    </item>
    <item>
      <title>Re: removing terms from a model following a designed experment</title>
      <link>https://community.jmp.com/t5/Discussions/removing-terms-from-a-model-following-a-designed-experment/m-p/703293#M88759</link>
      <description>&lt;P&gt;Why 4 replicates? &amp;nbsp;What was changing within and between replicate? &amp;nbsp;Does the RMSE look reasonable and comparable to the variation you see in the process? &amp;nbsp;Is the process stable? &amp;nbsp;&lt;/P&gt;
&lt;P&gt;&lt;FONT size="3" color="#000000"&gt;The commonest of defects in DOE are (paraphrased from Daniel):&lt;/FONT&gt;&lt;/P&gt;
&lt;OL&gt;
&lt;LI&gt;Oversaturation: too many effects for the number of treatments&lt;/LI&gt;
&lt;LI&gt;Overconservativeness: too many observations for the desired estimates&lt;/LI&gt;
&lt;LI&gt;Failure to study the data for bad values&lt;/LI&gt;
&lt;LI&gt;Failure to take into account all of the aliasing&lt;/LI&gt;
&lt;LI&gt;Imprecision due to misunderstanding the error variance.&lt;/LI&gt;
&lt;/OL&gt;
&lt;P&gt;Point 5 is important, particularly if you are relying on p-values for assessing significance.&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;Victor's advice regarding use of other statistics to determine an appropriate model are worth consideration as well.&lt;/P&gt;</description>
      <pubDate>Mon, 27 Nov 2023 17:20:52 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/removing-terms-from-a-model-following-a-designed-experment/m-p/703293#M88759</guid>
      <dc:creator>statman</dc:creator>
      <dc:date>2023-11-27T17:20:52Z</dc:date>
    </item>
    <item>
      <title>Re: removing terms from a model following a designed experment</title>
      <link>https://community.jmp.com/t5/Discussions/removing-terms-from-a-model-following-a-designed-experment/m-p/703321#M88763</link>
      <description>&lt;P&gt;The experiment was "open sourced" where volunteer participant teams conducted four runs (randomly assigned treatments) from the master design matrix. I blocked for the team effect, as well as other nuisance factors such as the scenario (of which there were four). Even without these nuisance factors, I expected a team's score (the response variable) to vary despite the factors' effects. These nuisance factors were an unfortunate compromise between "realism" with human subjects and a more tightly controlled experiment.&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Mon, 27 Nov 2023 18:20:51 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/removing-terms-from-a-model-following-a-designed-experment/m-p/703321#M88763</guid>
      <dc:creator>gchesterton</dc:creator>
      <dc:date>2023-11-27T18:20:51Z</dc:date>
    </item>
  </channel>
</rss>

