<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Re: Conceptual question about bagged predictions in Discussions</title>
    <link>https://community.jmp.com/t5/Discussions/Conceptual-question-about-bagged-predictions/m-p/457806#M70287</link>
    <description>&lt;P&gt;I believe the prediction confidence intervals for a multiple regression are derived via formulae that are derived from assumptions about the error structure and random sampling (ultimately derived from logic that underlies the Central Limit Theorem).&amp;nbsp; On the other hand, I believe the bagged predictions are derived from bootstrapping - a nonparametric empirical approach to deriving confidence intervals.&amp;nbsp; I also believe these two approaches are generally close, unless the underlying data has unusual distributions and/or insufficient sample sizes are used for the bootstrap.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;In the files I attached, I don't think any of these issues arise.&amp;nbsp; The mean predictions are almost identical from the 2 approaches, but the standard errors from the profiler are much much smaller than from the theoretically derived values.&amp;nbsp; This is what makes me think I am misunderstanding what the profiler standard errors mean.&amp;nbsp; Otherwise, why would anyone ever use the theoretically derived standard errors?&lt;/P&gt;</description>
    <pubDate>Wed, 02 Feb 2022 15:31:48 GMT</pubDate>
    <dc:creator>dale_lehman</dc:creator>
    <dc:date>2022-02-02T15:31:48Z</dc:date>
    <item>
      <title>Conceptual question about bagged predictions</title>
      <link>https://community.jmp.com/t5/Discussions/Conceptual-question-about-bagged-predictions/m-p/457314#M70248</link>
      <description>&lt;P&gt;I have struggled for a long while looking for uncertainty measures from machine learning models that are comparable to the standard errors that you routinely get from regression models.&amp;nbsp; Only recently I have become aware of some of the capabilities of the profiler - in particular, the bagged predictions.&amp;nbsp; But I don't really understand how to (or if I should) interpret those bagged predictions.&amp;nbsp; When I run a machine learning model (for example, a neural net) and I save the bagged predictions from the profiler, I get a Bagged mean, the standard error of the bagged mean, and the bagged standard deviation.&amp;nbsp; Comparing these with a regression model (multiple regression, for example), I've observed the following relationships:&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;UL&gt;&lt;LI&gt;The predicted bagged mean from the NN is very similar to the prediction formula from the multiple regression.&lt;/LI&gt;&lt;LI&gt;Mean prediction intervals from the multiple regression model are much narrower than the individual prediction intervals as expected (in the example I am looking at, the standard error for the mean prediction is about 1/10 the size of the standard error of the individual predictions).&lt;/LI&gt;&lt;LI&gt;The standard error of the bagged mean from the NN is much smaller than the bagged standard deviation (about 1/10 the size in the example I am looking at).&lt;/LI&gt;&lt;/UL&gt;&lt;P&gt;These observations tempt me to think of the standard error of the bagged mean from the NN as analogous to the standard error of the mean predictions from the regression model.&amp;nbsp; Similarly, the bagged standard deviation may be similar to the standard error of the individual predictions from the regression model.&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;However, the standard errors from the NN and the regression models do not resemble each other at all!&amp;nbsp; So, my question is whether my interpretation makes any sense - or, exactly how can the standard errors from the bagged mean be interpreted or used.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Thanks in advance for any insights.&amp;nbsp; I am attached an concrete example in case it helps with my question (this is the validation data set from my modeling example - with the predictions from the multiple regression model and NN included).&lt;/P&gt;</description>
      <pubDate>Fri, 09 Jun 2023 00:45:03 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/Conceptual-question-about-bagged-predictions/m-p/457314#M70248</guid>
      <dc:creator>dale_lehman</dc:creator>
      <dc:date>2023-06-09T00:45:03Z</dc:date>
    </item>
    <item>
      <title>Re: Conceptual question about bagged predictions</title>
      <link>https://community.jmp.com/t5/Discussions/Conceptual-question-about-bagged-predictions/m-p/457383#M70252</link>
      <description>&lt;P&gt;Addendum - in the file I attached, I (over)estimated the standard errors from the multiple regression.&amp;nbsp; I estimated them by using the saved 95% confidence intervals.&amp;nbsp; Attached is a revised version where I saved the standard errors directly.&amp;nbsp; My question remains, as the standard errors from the NN seem to be less than half as large as from the multiple regression (on average) with little apparent correlation within the validation set.&amp;nbsp; Both models fit the data quite well.&amp;nbsp; So, I am wondering if the interpretation of the standard errors of the bagged mean are appropriately interpreted as standard errors associated with the predictions in the same way that the standard errors of the predictions from the multiple regression are interpreted.&lt;/P&gt;</description>
      <pubDate>Mon, 31 Jan 2022 21:29:29 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/Conceptual-question-about-bagged-predictions/m-p/457383#M70252</guid>
      <dc:creator>dale_lehman</dc:creator>
      <dc:date>2022-01-31T21:29:29Z</dc:date>
    </item>
    <item>
      <title>Re: Conceptual question about bagged predictions</title>
      <link>https://community.jmp.com/t5/Discussions/Conceptual-question-about-bagged-predictions/m-p/457397#M70255</link>
      <description>&lt;P&gt;Yes.&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;The Neural platform uses a complex model, even with only a handful of hidden nodes. Imagine adding lots of terms to the regression model. What happens to the RMSE? What, in turn, happens to the SEs? Is your regression model complexity comparable to that of the NN?&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;I am not suggesting that you are misusing the NN or other ML methods. Just remember that they are ALL about prediction, NOT AT ALL about inference. So uncertainty in estimates is unimportant, accuracy / reproducibility / generalization are everything. So prediction models (not explanatory models) use measures of the total MSE (bias + variance) to select the model, and the 'honest assessment,' or cross-validation, to confirm the model that was selected.&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;But the main point is that there is no reason for the SEs from different models to be the same, even if their mean response predictions are similar. This difference might give one model an advantage in your application.&lt;/P&gt;</description>
      <pubDate>Mon, 31 Jan 2022 22:22:15 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/Conceptual-question-about-bagged-predictions/m-p/457397#M70255</guid>
      <dc:creator>Mark_Bailey</dc:creator>
      <dc:date>2022-01-31T22:22:15Z</dc:date>
    </item>
    <item>
      <title>Re: Conceptual question about bagged predictions</title>
      <link>https://community.jmp.com/t5/Discussions/Conceptual-question-about-bagged-predictions/m-p/457400#M70257</link>
      <description>&lt;P&gt;Mark, thanks.&amp;nbsp; That helps somewhat.&amp;nbsp; It might explain why the NN gives smaller standard errors (though I'm still surprised at the size of the difference when both models have such good fits to the data).&amp;nbsp; But it really doesn't seem to explain why there is almost no correlation in the standard errors associated with each prediction.&amp;nbsp; The data I posted shows virtually no meaningful relationship between each observation's standard error of prediction between the two models.&amp;nbsp; Now, for a multiple regression model I have some sense of what determines the standard errors associated with different observations - but for the NN, I really don't.&amp;nbsp; Perhaps that is the reason they are not related to each other?&amp;nbsp; Is this a dimension related to the lack of interpretability of NN models?&lt;/P&gt;</description>
      <pubDate>Tue, 01 Feb 2022 00:23:43 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/Conceptual-question-about-bagged-predictions/m-p/457400#M70257</guid>
      <dc:creator>dale_lehman</dc:creator>
      <dc:date>2022-02-01T00:23:43Z</dc:date>
    </item>
    <item>
      <title>Re: Conceptual question about bagged predictions</title>
      <link>https://community.jmp.com/t5/Discussions/Conceptual-question-about-bagged-predictions/m-p/457602#M70263</link>
      <description>&lt;P&gt;The NN is an ensemble model that is highly non-linear. Compare that with your regression model that might have second-order terms. The standard errors are very different for these two kinds of models.&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;I personally disagree with the notion that NN are not interpretable or that linear regression models are interpretable. (I think it is silly.) The effect of predictor X is a linear combination of all the terms that include it. For example, it is nonsense to talk about a 'quadratic effect.' There is only a quadratic term in the model. So how do we interpret the effect of X when it appears in the model as X + X*X2 + X*X + X*X*X? We are just used to thinking in these terms - we had many years of exposure to it and time to think about it. A NN is more of the same (linear predictor) put through a non-linear activation function, and added to more of the same for each node. We just have to think harder.&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;No, we don't. We have a profiler that works with any kind of function (model).&lt;/P&gt;</description>
      <pubDate>Tue, 01 Feb 2022 17:02:44 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/Conceptual-question-about-bagged-predictions/m-p/457602#M70263</guid>
      <dc:creator>Mark_Bailey</dc:creator>
      <dc:date>2022-02-01T17:02:44Z</dc:date>
    </item>
    <item>
      <title>Re: Conceptual question about bagged predictions</title>
      <link>https://community.jmp.com/t5/Discussions/Conceptual-question-about-bagged-predictions/m-p/457633#M70269</link>
      <description>&lt;P&gt;I'm following what you say - but I think my question has become something different.&amp;nbsp; Let's leave NN out of it.&amp;nbsp; Since the profiler is available from the multiple regression platform as well, I did some experimenting to see how the standard errors of the bagged predictions compare with those saved from the regression model.&amp;nbsp; They are correlated, though far from perfectly.&amp;nbsp; The individual prediction standard deviation is much larger than the mean prediction standard deviation, as it should be, and this also applies to the two standard errors you get when you save the bagged means.&amp;nbsp; However, what surprises me and I don't understand, is why the standard errors from saving the bagged predictions (either the individual or mean version) are an order of magnitude smaller than the standard errors from the regression model.&amp;nbsp; My understanding (which could be wrong) is that the standard errors of the predictions are theoretically derived in the regression model and are the result of bootstrapping in the bagged predictions.&amp;nbsp; In theory, those should be similar (at least with enough bootstrap samples - I used 100 and 1000 and both give similar results) - but they are not.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;So, it appears that the standard errors from the profiler are qualitatively different than the standard errors from the regression model.&amp;nbsp; Why is that the case?&lt;/P&gt;</description>
      <pubDate>Tue, 01 Feb 2022 19:39:03 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/Conceptual-question-about-bagged-predictions/m-p/457633#M70269</guid>
      <dc:creator>dale_lehman</dc:creator>
      <dc:date>2022-02-01T19:39:03Z</dc:date>
    </item>
    <item>
      <title>Re: Conceptual question about bagged predictions</title>
      <link>https://community.jmp.com/t5/Discussions/Conceptual-question-about-bagged-predictions/m-p/457634#M70270</link>
      <description>&lt;P&gt;Trying to be more complete, attached is a simple little file with a single predictor.&amp;nbsp; I built the regression model and saved the predictions and standard errors (both mean and individual).&amp;nbsp; I used the profiler to save the bagged predictions.&amp;nbsp; There are 3 xy plots showing the equivalence (perfect in this case) of the predicted values from the model and from the profiler.&amp;nbsp; The standard error plots show strong correlations, but also show that the standard errors from the profiler are considerably smaller than from the model.&amp;nbsp; I strongly suspect the standard errors from the bagging procedure mean something different.&amp;nbsp; I hope you can clarify this for me.&lt;/P&gt;</description>
      <pubDate>Tue, 01 Feb 2022 19:53:45 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/Conceptual-question-about-bagged-predictions/m-p/457634#M70270</guid>
      <dc:creator>dale_lehman</dc:creator>
      <dc:date>2022-02-01T19:53:45Z</dc:date>
    </item>
    <item>
      <title>Re: Conceptual question about bagged predictions</title>
      <link>https://community.jmp.com/t5/Discussions/Conceptual-question-about-bagged-predictions/m-p/457805#M70286</link>
      <description>&lt;P&gt;"&lt;SPAN&gt;My understanding (which could be wrong) is that the standard errors of the predictions are theoretically derived in the regression model and are the result of bootstrapping in the bagged predictions.&amp;nbsp; In theory, those should be similar (at least with enough bootstrap samples - I used 100 and 1000 and both give similar results) - but they are not.&lt;/SPAN&gt;"&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;Which theory are you referring to?&lt;/P&gt;</description>
      <pubDate>Wed, 02 Feb 2022 15:18:04 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/Conceptual-question-about-bagged-predictions/m-p/457805#M70286</guid>
      <dc:creator>Mark_Bailey</dc:creator>
      <dc:date>2022-02-02T15:18:04Z</dc:date>
    </item>
    <item>
      <title>Re: Conceptual question about bagged predictions</title>
      <link>https://community.jmp.com/t5/Discussions/Conceptual-question-about-bagged-predictions/m-p/457806#M70287</link>
      <description>&lt;P&gt;I believe the prediction confidence intervals for a multiple regression are derived via formulae that are derived from assumptions about the error structure and random sampling (ultimately derived from logic that underlies the Central Limit Theorem).&amp;nbsp; On the other hand, I believe the bagged predictions are derived from bootstrapping - a nonparametric empirical approach to deriving confidence intervals.&amp;nbsp; I also believe these two approaches are generally close, unless the underlying data has unusual distributions and/or insufficient sample sizes are used for the bootstrap.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;In the files I attached, I don't think any of these issues arise.&amp;nbsp; The mean predictions are almost identical from the 2 approaches, but the standard errors from the profiler are much much smaller than from the theoretically derived values.&amp;nbsp; This is what makes me think I am misunderstanding what the profiler standard errors mean.&amp;nbsp; Otherwise, why would anyone ever use the theoretically derived standard errors?&lt;/P&gt;</description>
      <pubDate>Wed, 02 Feb 2022 15:31:48 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/Conceptual-question-about-bagged-predictions/m-p/457806#M70287</guid>
      <dc:creator>dale_lehman</dc:creator>
      <dc:date>2022-02-02T15:31:48Z</dc:date>
    </item>
    <item>
      <title>Re: Conceptual question about bagged predictions</title>
      <link>https://community.jmp.com/t5/Discussions/Conceptual-question-about-bagged-predictions/m-p/457812#M70290</link>
      <description>&lt;P&gt;"&lt;SPAN&gt;I believe the prediction confidence intervals for a multiple regression are derived via formulae that are derived from assumptions about the error structure and random sampling (ultimately derived from logic that underlies the Central Limit Theorem).&amp;nbsp; On the other hand, I believe the bagged predictions are derived from bootstrapping - a nonparametric empirical approach to deriving confidence intervals.&amp;nbsp; I also believe these two approaches are generally close, unless the underlying data has unusual distributions and/or insufficient sample sizes are used for the bootstrap.&lt;/SPAN&gt;"&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;Yes! That is, if you use the theoretical expression for the CI and the bootstrap CI for the SAME linear regression model, the CI ESTIMATES should agree.&lt;/P&gt;</description>
      <pubDate>Wed, 02 Feb 2022 16:14:14 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/Conceptual-question-about-bagged-predictions/m-p/457812#M70290</guid>
      <dc:creator>Mark_Bailey</dc:creator>
      <dc:date>2022-02-02T16:14:14Z</dc:date>
    </item>
    <item>
      <title>Re: Conceptual question about bagged predictions</title>
      <link>https://community.jmp.com/t5/Discussions/Conceptual-question-about-bagged-predictions/m-p/457816#M70291</link>
      <description>&lt;P&gt;Mark&lt;/P&gt;&lt;P&gt;Then I would ask you to look at the last example dataset I posted.&amp;nbsp; The confidence intervals (standard errors) are not even close to agreeing.&amp;nbsp; That is for a simple linear regression and comparing the prediction standard errors with those that come from saving the bagged predictions from the profiler.&amp;nbsp; The latter are much much smaller than the former.&amp;nbsp; That is why I am confused.&lt;/P&gt;</description>
      <pubDate>Wed, 02 Feb 2022 16:19:42 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/Conceptual-question-about-bagged-predictions/m-p/457816#M70291</guid>
      <dc:creator>dale_lehman</dc:creator>
      <dc:date>2022-02-02T16:19:42Z</dc:date>
    </item>
    <item>
      <title>Re: Conceptual question about bagged predictions</title>
      <link>https://community.jmp.com/t5/Discussions/Conceptual-question-about-bagged-predictions/m-p/457843#M70293</link>
      <description>&lt;P&gt;Let's review how the bagging works in the Profiler. Note that it is based on the fitted model. It uses the fitted model and resamples the data to inflate the data set or sample size but does not alter the model. Here is an excerpt from the JMP Help that covers &lt;A href="https://www.jmp.com/support/help/en/16.1/#page/jmp/bagging.shtml#ww484393" target="_self"&gt;Profiler bagging&lt;/A&gt;:&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;"Bagging automatically creates new columns in the original data table. All&lt;SPAN&gt;&amp;nbsp;&lt;/SPAN&gt;&lt;SPAN class="EquationVariables"&gt;M&lt;/SPAN&gt;&lt;SPAN&gt;&amp;nbsp;&lt;/SPAN&gt;sets of bagged predictions are saved as hidden columns. The final prediction is saved in a column named “Pred Formula &amp;lt;colname&amp;gt; Bagged Mean”. The standard deviation of the final prediction is saved in a column named “&amp;lt;colname&amp;gt; Bagged Std Dev”. The standard error of the bagged mean is saved in a column named “StdError &amp;lt;colname&amp;gt; Bagged Mean.” The standard error is the standard deviation divided by&lt;SPAN&gt;&amp;nbsp;Sqrt( M-1 ).&amp;nbsp;&lt;/SPAN&gt;Here, &amp;lt;colname&amp;gt; identifies the column in the report that was bagged.&lt;/P&gt;
&lt;P class="body"&gt;The standard error gives insight about the precision of the prediction. A very small standard error indicates a precise prediction for that observation. For more information about bagging, see Hastie et al. (&lt;SPAN class="link"&gt;&lt;A title="References" href="https://www.jmp.com/support/help/en/16.1/jmp/references-4.shtml#ww130941" target="_blank"&gt;2009&lt;/A&gt;&lt;/SPAN&gt;)."&lt;/P&gt;
&lt;P class="body"&gt;&amp;nbsp;&lt;/P&gt;
&lt;P class="body"&gt;So you are not comparing the SEs from a regression analysis and NN. You are comparing SEs from any model and the SEs from bagging in the Profiler with the same model. The difference will be a factor of Sqrt( M-1 ).&lt;/P&gt;</description>
      <pubDate>Wed, 02 Feb 2022 16:52:40 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/Conceptual-question-about-bagged-predictions/m-p/457843#M70293</guid>
      <dc:creator>Mark_Bailey</dc:creator>
      <dc:date>2022-02-02T16:52:40Z</dc:date>
    </item>
    <item>
      <title>Re: Conceptual question about bagged predictions</title>
      <link>https://community.jmp.com/t5/Discussions/Conceptual-question-about-bagged-predictions/m-p/457845#M70295</link>
      <description>&lt;P&gt;We are almost on the same page.&amp;nbsp; Ignoring NN for the moment, and using the simple regression example I provided, the standard error from the bagged mean is about 10% of that from the regression model:&amp;nbsp; this matches the square root of M-1 (M=100 here).&amp;nbsp; That is true for the mean predictions - for the individual predictions it is about 20% (I'm not sure why that sq rt (M-1) applies on the former but not the latter, but I don't think that is very important).&amp;nbsp; So, the question is:&amp;nbsp; if I want a confidence interval for the mean prediction, which standard error do I use?&amp;nbsp; The difference is an order of magnitude!&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Assuming for the moment that I should use the standard error from the regression model (which is the larger of the two) - and that is what I suspect is the case - then it raises the question of what the bagged predictions are good for.&amp;nbsp; Here, I think the NN (and other machine learning models) comes in - there is no standard error from many of these models without some type of empirically derived one, such as provided by the profiler.&amp;nbsp; So, it would seem very useful to use that standard error from the bagged predictions to construct confidence intervals for those predictions.&amp;nbsp; However, from the regression example, I am wondering if this might underestimate the degree uncertainty by an order of magnitude - which would not be so useful.&lt;/P&gt;</description>
      <pubDate>Wed, 02 Feb 2022 17:47:00 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/Conceptual-question-about-bagged-predictions/m-p/457845#M70295</guid>
      <dc:creator>dale_lehman</dc:creator>
      <dc:date>2022-02-02T17:47:00Z</dc:date>
    </item>
    <item>
      <title>Re: Conceptual question about bagged predictions</title>
      <link>https://community.jmp.com/t5/Discussions/Conceptual-question-about-bagged-predictions/m-p/457847#M70296</link>
      <description>&lt;P&gt;"&lt;SPAN&gt;So, the question is:&amp;nbsp; if I want a confidence interval for the mean prediction, which standard error do I use?&amp;nbsp; The difference is an order of magnitude!&lt;/SPAN&gt;" Use the SE from the model that was used to predict.&amp;nbsp;If you use the saved regression model, then use the saved SE or CI. If you are using the bagged predictions from the Profiler, then use the bagged SE. Here is the result using :weight versus :height in the Big Class data set. The highlighted pairs of prediction and SE data columns would be used together.&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&lt;span class="lia-inline-image-display-wrapper lia-image-align-inline" image-alt="bagged.PNG" style="width: 616px;"&gt;&lt;img src="https://community.jmp.com/t5/image/serverpage/image-id/39671iF35911A3C0BBF3EF/image-size/large?v=v2&amp;amp;px=999" role="button" title="bagged.PNG" alt="bagged.PNG" /&gt;&lt;/span&gt;&lt;/P&gt;</description>
      <pubDate>Wed, 02 Feb 2022 18:02:23 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/Conceptual-question-about-bagged-predictions/m-p/457847#M70296</guid>
      <dc:creator>Mark_Bailey</dc:creator>
      <dc:date>2022-02-02T18:02:23Z</dc:date>
    </item>
    <item>
      <title>Re: Conceptual question about bagged predictions</title>
      <link>https://community.jmp.com/t5/Discussions/Conceptual-question-about-bagged-predictions/m-p/457867#M70298</link>
      <description>&lt;P&gt;Surely that can't be right!&amp;nbsp; Your example looks just like mine - the means are almost identical but the standard deviations differ by a factor of 10 (due to the sq rt (M-1) depending on which sets of columns you use.&amp;nbsp; So, while I see the logic of pairing the mean prediction with its associated standard deviation (either from the model or from the bagging), the practical effect is to have roughly the same mean predictions, but one confidence interval ends up being 10% as wide as the other.&amp;nbsp; Which one is the appropriate measure of variability for the mean prediction?&amp;nbsp; It can't be both - unless they answer different questions.&amp;nbsp; And, if that is the case, can you tell me what question each answers?&lt;/P&gt;</description>
      <pubDate>Wed, 02 Feb 2022 18:16:26 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/Conceptual-question-about-bagged-predictions/m-p/457867#M70298</guid>
      <dc:creator>dale_lehman</dc:creator>
      <dc:date>2022-02-02T18:16:26Z</dc:date>
    </item>
    <item>
      <title>Re: Conceptual question about bagged predictions</title>
      <link>https://community.jmp.com/t5/Discussions/Conceptual-question-about-bagged-predictions/m-p/457890#M70302</link>
      <description>&lt;P&gt;"&lt;SPAN&gt;Surely that can't be right&lt;/SPAN&gt;"&lt;/P&gt;
&lt;P&gt;OK, you got me. It is all made up, faked. I was just seeing how far I could string you along.&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;(Serious discussion resumes...)&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;"&lt;SPAN&gt;Which one is the appropriate measure of variability for the mean prediction?&lt;/SPAN&gt;"&lt;/P&gt;
&lt;P&gt;The one that was estimated for prediction you will use. The pairing thing...&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;Bagging is more about predictive modeling than explanatory modeling, as I explained. Bagging decreases the uncertainty in the prediction.&amp;nbsp;The use of bagging in this case relies on your belief in the quality and validity of the data and the model. it is not cheating. If the model fails, it is because of a problem with the quality or validity of the data or the model.&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;"&lt;SPAN&gt;It can't be both - unless they answer different questions.&lt;/SPAN&gt;"&lt;/P&gt;
&lt;P&gt;That is exactly what I have been saying.&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;"&lt;SPAN&gt;And, if that is the case, can you tell me what question each answers?&lt;/SPAN&gt;"&lt;/P&gt;
&lt;P&gt;I can. I actually did: again, the first pair answers the question about the uncertainty in the prediction of the original model (e.g., linear regression, neural network, partition). The second pair answers the question about the uncertainty in the prediction using bagging.&lt;/P&gt;</description>
      <pubDate>Wed, 02 Feb 2022 19:36:21 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/Conceptual-question-about-bagged-predictions/m-p/457890#M70302</guid>
      <dc:creator>Mark_Bailey</dc:creator>
      <dc:date>2022-02-02T19:36:21Z</dc:date>
    </item>
    <item>
      <title>Re: Conceptual question about bagged predictions</title>
      <link>https://community.jmp.com/t5/Discussions/Conceptual-question-about-bagged-predictions/m-p/457892#M70304</link>
      <description>&lt;P&gt;We are converging.&lt;/P&gt;&lt;P&gt;"And, if that is the case, can you tell me what question each answers?"&lt;/P&gt;&lt;P&gt;I can. I actually did: again, the first pair answers the question about the uncertainty in the prediction of the original model (e.g., linear regression, neural network, partition). The second pair answers the question about the uncertainty in the prediction using bagging.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;This says that I have more uncertainty about my model predictions than I have about the bagged predictions - a lot more.&amp;nbsp; Why would anyone use the model predictions and their confidence intervals then?&amp;nbsp; I realize that a smaller standard error is not always good - only if the underlying model is good.&amp;nbsp; But in the case we are looking at, the same model underlies both measures and the mean predictions are almost identical.&amp;nbsp; So, why would I choose the much wider confidence interval?&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;I suppose there is the issue of coverage - the narrow interval might not provide enough coverage of the true value.&amp;nbsp; I will try some simulations to see if I can shed any light on that - but do you know of any references that speak to the accuracy of the two standard error measures relative to each other?&lt;/P&gt;</description>
      <pubDate>Wed, 02 Feb 2022 19:57:23 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/Conceptual-question-about-bagged-predictions/m-p/457892#M70304</guid>
      <dc:creator>dale_lehman</dc:creator>
      <dc:date>2022-02-02T19:57:23Z</dc:date>
    </item>
    <item>
      <title>Re: Conceptual question about bagged predictions</title>
      <link>https://community.jmp.com/t5/Discussions/Conceptual-question-about-bagged-predictions/m-p/457893#M70305</link>
      <description>&lt;P&gt;Attached is a simulated example.&amp;nbsp; The untitled dataset contains simulated data and a regression model based on a random sample of 100 of the 1000 rows.&amp;nbsp; I saved the standard errors from the model and from the bagged predictions.&amp;nbsp; The subset of untitled data set then contains the 900 rows not in the random sample.&amp;nbsp; Using approximate 95% confidence intervals (2 standard errors around the corresponding mean prediction, using the standard errors for the individual predictions), the coverage of the actual Y value was 866 out of 900 rows (around 95%) for the model predictions, but only 222 out of 900 for the bagged predictions.&amp;nbsp; I couldn't figure out how to generate a comparison of coverage of mean confidence intervals since I only have a single Y observation on each row.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Given how extreme the results are, this suggests to me that the standard errors from the model (at least for this well behaved model) are accurate measures of the uncertainty in the predictions.&amp;nbsp; But the bagged standard errors are too small.&amp;nbsp; Given the simplicity of this example, it sure seems like I wouldn't want to rely on the standard errors from the bagged predictions.&amp;nbsp; Do you think that is a reasonable conclusion here?&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Now, to the real potential uses.&amp;nbsp; If I run a classification model using NN, random forests, boosted trees, etc., one shortcoming compared with logistic regression is that these machine learning models do not provide a measure of uncertainty in the predictions (without invoking another procedure such as conformal prediction, which I have been playing with).&amp;nbsp; The profiler could readily provide me with bagged predictions of the mean probabilities of classifications along with their standard errors.&amp;nbsp; As useful as that would be, I am inclined to say that I can't really use those standard errors to represent the uncertainty in these machine learning models.&amp;nbsp; Is that correct?&amp;nbsp; Perhaps a more general question is, what can I use the bagged standard errors for?&lt;/P&gt;</description>
      <pubDate>Wed, 02 Feb 2022 20:24:55 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/Conceptual-question-about-bagged-predictions/m-p/457893#M70305</guid>
      <dc:creator>dale_lehman</dc:creator>
      <dc:date>2022-02-02T20:24:55Z</dc:date>
    </item>
    <item>
      <title>Re: Conceptual question about bagged predictions</title>
      <link>https://community.jmp.com/t5/Discussions/Conceptual-question-about-bagged-predictions/m-p/457894#M70306</link>
      <description>&lt;P&gt;"&lt;SPAN&gt;So, why would I choose the much wider confidence interval?&lt;/SPAN&gt;"&lt;/P&gt;
&lt;P&gt;Because you question the quality or the validity of the data or the model.&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;I would likely not use bagging and its predictions with a screening experiment because the model is likely biased. I would likely not use bagging and its predictions with a model based on a small sample of observational data.&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;Let me be clear. Bagging is valid. It is not cheating. But it is not always appropriate or beneficial.&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;Unfortunately, bagging in the profiler is something JMP developed. I do not have external references.&lt;/P&gt;</description>
      <pubDate>Wed, 02 Feb 2022 20:37:39 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/Conceptual-question-about-bagged-predictions/m-p/457894#M70306</guid>
      <dc:creator>Mark_Bailey</dc:creator>
      <dc:date>2022-02-02T20:37:39Z</dc:date>
    </item>
    <item>
      <title>Re: Conceptual question about bagged predictions</title>
      <link>https://community.jmp.com/t5/Discussions/Conceptual-question-about-bagged-predictions/m-p/457895#M70307</link>
      <description>&lt;P&gt;In my simulated example, I have no reason to question the quality of the data or appropriateness of the model.&amp;nbsp; It is not observational data, nor is it a small sample (true, n=100 out of a population of 1000 is not large, but the confidence interval coverage is so disparate that I think the example shows us something).&amp;nbsp; I won't say bagging is cheating (a loaded term).&amp;nbsp; But I don't feel like I can trust the standard errors that it provides even for my simulated case.&amp;nbsp; I am very reluctant to use it for a case where the data and model are more suspect.&amp;nbsp; Can you provide any guidance for where it can be used?&amp;nbsp; I don't mean to be antagonistic:&amp;nbsp; I love JMP and love the profiler, I'm just trying to see whether the confidence intervals it can provide are useful.&lt;/P&gt;</description>
      <pubDate>Wed, 02 Feb 2022 21:13:14 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/Conceptual-question-about-bagged-predictions/m-p/457895#M70307</guid>
      <dc:creator>dale_lehman</dc:creator>
      <dc:date>2022-02-02T21:13:14Z</dc:date>
    </item>
  </channel>
</rss>

