<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Difference in results between Generalized RSquare and RSquare in Discussions</title>
    <link>https://community.jmp.com/t5/Discussions/Difference-in-results-between-Generalized-RSquare-and-RSquare/m-p/336638#M58404</link>
    <description>&lt;P&gt;Hi everyone,&amp;nbsp;&lt;/P&gt;&lt;P&gt;I tried to find an answer to my question in the community or in JMP help, but I couldn't. So, I would really appreciate if anyone could point me in the right direction here or let me know what is the basic concept I am missing.&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;I am fitting a model to my response (dependent variable - DV), which is continuous, using 3 independent variables (IV) [2 categorical and 1 continuous] and the interaction between 2 of them (the other one I am just using the main effect). I first used Least Squares for that, as you can see in the screenshot below.&lt;/P&gt;&lt;P&gt;&lt;span class="lia-inline-image-display-wrapper lia-image-align-inline" image-alt="AnnaPaula_0-1606154145936.png" style="width: 400px;"&gt;&lt;img src="https://community.jmp.com/t5/image/serverpage/image-id/28557i207D9A1000D031B9/image-size/medium?v=v2&amp;amp;px=400" role="button" title="AnnaPaula_0-1606154145936.png" alt="AnnaPaula_0-1606154145936.png" /&gt;&lt;/span&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;However, my residual by predicted plot does not look very good (I have some outliers).&amp;nbsp;No matter which other factors I tried to add to the model - or transformation - I was not able to improve my residuals plot. The best I could get was using a change of my response, which is the one I am already showing above.&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Next, I thought about using the Penalized Regression Method (more specifically Ridge) through the Generalized Regression. I read it can be more robust to outliers (though it will introduce some bias).&lt;/P&gt;&lt;P&gt;I started running the Generalized Regression with Standard Least Squares.&lt;/P&gt;&lt;P&gt;According to some videos I have seen, this should produce the same results as the Least Square regression.&lt;/P&gt;&lt;P&gt;&lt;STRONG&gt;Also, according to JMP help (&lt;A href="https://www.jmp.com/support/help/en/15.2/index.shtml#page/jmp/training-and-validation-measures-of-fit.shtml" target="_blank" rel="noopener"&gt;https://www.jmp.com/support/help/en/15.2/index.shtml#page/jmp/training-and-validation-measures-of-fit.shtml&lt;/A&gt;) the Generalized RSquare provided should be between 0 and 1 and it will&amp;nbsp;simplify to the traditional RSquare for continuous normal responses in the standard least-squares setting. However, as you can see from the screenshot below, although the Generalized RSquare matches the RSquare for the Training set, it does not match for the Validation set (and it is negative). The other statistics (e.g. RASE) do match both sets (Training and Validation).&lt;/STRONG&gt;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;Any ideas about what may be happening here?&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;&amp;nbsp;&lt;span class="lia-inline-image-display-wrapper lia-image-align-inline" image-alt="image.png" style="width: 429px;"&gt;&lt;img src="https://community.jmp.com/t5/image/serverpage/image-id/28561i36B54CA2865DA9BA/image-dimensions/429x387?v=v2" width="429" height="387" role="button" title="image.png" alt="image.png" /&gt;&lt;/span&gt;&lt;/SPAN&gt;&lt;/P&gt;&lt;DIV class="mceNonEditable lia-copypaste-placeholder"&gt;&amp;nbsp;&lt;/DIV&gt;&lt;DIV class="mceNonEditable lia-copypaste-placeholder"&gt;&amp;nbsp;&lt;/DIV&gt;&lt;P&gt;Thank you in advance for any ideas and hints!&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;PS:&amp;nbsp;For everything I am doing in this post, I used the FitModel platform in JMP.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
    <pubDate>Fri, 09 Jun 2023 00:25:14 GMT</pubDate>
    <dc:creator>AnnaPaula</dc:creator>
    <dc:date>2023-06-09T00:25:14Z</dc:date>
    <item>
      <title>Difference in results between Generalized RSquare and RSquare</title>
      <link>https://community.jmp.com/t5/Discussions/Difference-in-results-between-Generalized-RSquare-and-RSquare/m-p/336638#M58404</link>
      <description>&lt;P&gt;Hi everyone,&amp;nbsp;&lt;/P&gt;&lt;P&gt;I tried to find an answer to my question in the community or in JMP help, but I couldn't. So, I would really appreciate if anyone could point me in the right direction here or let me know what is the basic concept I am missing.&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;I am fitting a model to my response (dependent variable - DV), which is continuous, using 3 independent variables (IV) [2 categorical and 1 continuous] and the interaction between 2 of them (the other one I am just using the main effect). I first used Least Squares for that, as you can see in the screenshot below.&lt;/P&gt;&lt;P&gt;&lt;span class="lia-inline-image-display-wrapper lia-image-align-inline" image-alt="AnnaPaula_0-1606154145936.png" style="width: 400px;"&gt;&lt;img src="https://community.jmp.com/t5/image/serverpage/image-id/28557i207D9A1000D031B9/image-size/medium?v=v2&amp;amp;px=400" role="button" title="AnnaPaula_0-1606154145936.png" alt="AnnaPaula_0-1606154145936.png" /&gt;&lt;/span&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;However, my residual by predicted plot does not look very good (I have some outliers).&amp;nbsp;No matter which other factors I tried to add to the model - or transformation - I was not able to improve my residuals plot. The best I could get was using a change of my response, which is the one I am already showing above.&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Next, I thought about using the Penalized Regression Method (more specifically Ridge) through the Generalized Regression. I read it can be more robust to outliers (though it will introduce some bias).&lt;/P&gt;&lt;P&gt;I started running the Generalized Regression with Standard Least Squares.&lt;/P&gt;&lt;P&gt;According to some videos I have seen, this should produce the same results as the Least Square regression.&lt;/P&gt;&lt;P&gt;&lt;STRONG&gt;Also, according to JMP help (&lt;A href="https://www.jmp.com/support/help/en/15.2/index.shtml#page/jmp/training-and-validation-measures-of-fit.shtml" target="_blank" rel="noopener"&gt;https://www.jmp.com/support/help/en/15.2/index.shtml#page/jmp/training-and-validation-measures-of-fit.shtml&lt;/A&gt;) the Generalized RSquare provided should be between 0 and 1 and it will&amp;nbsp;simplify to the traditional RSquare for continuous normal responses in the standard least-squares setting. However, as you can see from the screenshot below, although the Generalized RSquare matches the RSquare for the Training set, it does not match for the Validation set (and it is negative). The other statistics (e.g. RASE) do match both sets (Training and Validation).&lt;/STRONG&gt;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;Any ideas about what may be happening here?&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;&amp;nbsp;&lt;span class="lia-inline-image-display-wrapper lia-image-align-inline" image-alt="image.png" style="width: 429px;"&gt;&lt;img src="https://community.jmp.com/t5/image/serverpage/image-id/28561i36B54CA2865DA9BA/image-dimensions/429x387?v=v2" width="429" height="387" role="button" title="image.png" alt="image.png" /&gt;&lt;/span&gt;&lt;/SPAN&gt;&lt;/P&gt;&lt;DIV class="mceNonEditable lia-copypaste-placeholder"&gt;&amp;nbsp;&lt;/DIV&gt;&lt;DIV class="mceNonEditable lia-copypaste-placeholder"&gt;&amp;nbsp;&lt;/DIV&gt;&lt;P&gt;Thank you in advance for any ideas and hints!&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;PS:&amp;nbsp;For everything I am doing in this post, I used the FitModel platform in JMP.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Fri, 09 Jun 2023 00:25:14 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/Difference-in-results-between-Generalized-RSquare-and-RSquare/m-p/336638#M58404</guid>
      <dc:creator>AnnaPaula</dc:creator>
      <dc:date>2023-06-09T00:25:14Z</dc:date>
    </item>
    <item>
      <title>Re: Difference in results between Generalized RSquare and RSquare</title>
      <link>https://community.jmp.com/t5/Discussions/Difference-in-results-between-Generalized-RSquare-and-RSquare/m-p/336681#M58407</link>
      <description>&lt;P&gt;Is the continuous predictor involved in the interaction term? Did you use the estimate for the omitted interaction to determine that it should be removed or did you leave it out to begin with?&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;What transformations of the response did you try?&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;Penalized regression and least squares regression do not usually lead to the same estimates of the model parameters, so the R square would not be the same. Your picture shows the OLS result, not the Ridge Regression.&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;The training and validation results should be similar with the best unbiased model, but not the same. How did you determine the hold out set for validation?&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;The negative R square is possible.&lt;/P&gt;</description>
      <pubDate>Mon, 23 Nov 2020 18:26:44 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/Difference-in-results-between-Generalized-RSquare-and-RSquare/m-p/336681#M58407</guid>
      <dc:creator>Mark_Bailey</dc:creator>
      <dc:date>2020-11-23T18:26:44Z</dc:date>
    </item>
    <item>
      <title>Re: Difference in results between Generalized RSquare and RSquare</title>
      <link>https://community.jmp.com/t5/Discussions/Difference-in-results-between-Generalized-RSquare-and-RSquare/m-p/336689#M58409</link>
      <description>&lt;P&gt;Hi&amp;nbsp;&lt;a href="https://community.jmp.com/t5/user/viewprofilepage/user-id/21986"&gt;@AnnaPaula&lt;/a&gt;&amp;nbsp;,&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp; If I understand correctly, you have a data set that you have split according to a validation column of some kind. One thing to consider is whether or not you stratified the validation column on your response to make sure both the training and validation columns have similar distributions.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp; Since the R^2 is negative, to me that would indicate you're overfitting your data. Have you done some bootstrapping on the estimates to see if the terms you have in your model are really important? Do you have a theoretical basis for including mixed terms?&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp; Also, you might want to try Elastic Net since it's a combination of Ridge and Lasso. Another thing to consider is if your response distribution is really normal or if a different distribution is better suited to your data.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;DS&lt;/P&gt;</description>
      <pubDate>Mon, 23 Nov 2020 18:37:10 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/Difference-in-results-between-Generalized-RSquare-and-RSquare/m-p/336689#M58409</guid>
      <dc:creator>SDF1</dc:creator>
      <dc:date>2020-11-23T18:37:10Z</dc:date>
    </item>
    <item>
      <title>Re: Difference in results between Generalized RSquare and RSquare</title>
      <link>https://community.jmp.com/t5/Discussions/Difference-in-results-between-Generalized-RSquare-and-RSquare/m-p/336740#M58418</link>
      <description>&lt;P&gt;Sorry&amp;nbsp;&lt;a href="https://community.jmp.com/t5/user/viewprofilepage/user-id/5358"&gt;@Mark_Bailey&lt;/a&gt;, I think I was not very clear in my question.&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;As you stated I did not use Ridge (that was gonna be my next step). But first, I just used again the Least Squares method (through the Generalized Regression model), to check if the results would match with the results using LeastSquares directly from the FitModel window. Because my results did not match, I did not even end up trying Ridge (which would be my next step).&lt;/P&gt;&lt;P&gt;But I am curious &lt;STRONG&gt;why the results from LeastSquares through Generalized Regression are not matching the results through LeastSquares directly from FitModel&lt;/STRONG&gt;. Do you have any ideas?&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Answering your questions:&lt;/P&gt;&lt;P&gt;1. The continuous predictor is not involved in the interaction term.&lt;/P&gt;&lt;P&gt;2. I used stepwise and also used the estimates myself to determine whether the interaction effect should be removed&lt;/P&gt;&lt;P&gt;3. I tried Log, Logistic, Logit, and Reciprocal. With reciprocal, the model deteriorates a lot. With Logistic, there is a slight improvement in the statistics (R-square, RASE), but not in the residual plots. So nothing to justify the use of the transformation in my opinion.&amp;nbsp;&lt;/P&gt;&lt;P&gt;4. The training and validation were defined based on groups I have (these are computer experiments with different random seeds). I used the seeds to separate the data in groups and used that column as my holdout validation column.&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;5. My question related to negative was for &lt;STRONG&gt;Generalized R-square&lt;/STRONG&gt;, cause when I first read the JMP help, I thought it could only go between 0 and 1. But yes, I found the formula for that one and as we can see it can also go below 0:&amp;nbsp;&lt;/P&gt;&lt;TABLE border="1"&gt;&lt;TBODY&gt;&lt;TR&gt;&lt;TD&gt;Nagelkerke / Cragg &amp;amp; Uhler’s (Generalized RSquare)&lt;/TD&gt;&lt;TD&gt;&lt;DIV class="mceNonEditable lia-copypaste-placeholder"&gt;&amp;nbsp;&lt;/DIV&gt;&lt;P&gt;&lt;span class="lia-inline-image-display-wrapper lia-image-align-inline" image-alt="image.png" style="width: 237px;"&gt;&lt;img src="https://community.jmp.com/t5/image/serverpage/image-id/28567iE891AFBCC45888B8/image-size/large?v=v2&amp;amp;px=999" role="button" title="image.png" alt="image.png" /&gt;&lt;/span&gt;&lt;/P&gt;&lt;/TD&gt;&lt;/TR&gt;&lt;/TBODY&gt;&lt;/TABLE&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Thank you!&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Mon, 23 Nov 2020 21:59:11 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/Difference-in-results-between-Generalized-RSquare-and-RSquare/m-p/336740#M58418</guid>
      <dc:creator>AnnaPaula</dc:creator>
      <dc:date>2020-11-23T21:59:11Z</dc:date>
    </item>
    <item>
      <title>Re: Difference in results between Generalized RSquare and RSquare</title>
      <link>https://community.jmp.com/t5/Discussions/Difference-in-results-between-Generalized-RSquare-and-RSquare/m-p/336741#M58419</link>
      <description>&lt;P&gt;Hi&amp;nbsp;&lt;a href="https://community.jmp.com/t5/user/viewprofilepage/user-id/12549"&gt;@SDF1&lt;/a&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Thank you so much for your reply and your insights. They are very helpful and they align with a few things I was thinking.&amp;nbsp;&lt;/P&gt;&lt;P&gt;I will provide some extra information here in case you have more insights to give (I would really appreciate it).&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;1. The stratified response based on the validation columns do have (very) similar distributions. As I explained to Mark, the response was calculated based on computer experiments generated using different seeds (and I used the seeds to stratify into training and validation sets).&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;2. Let's assume I am overfitting, my main question is:&lt;/P&gt;&lt;P&gt;Why my RSquare for the training set is 0.9995 using LeastSquares directly from FitModel or through Generalized Regression, while my RSquare for the validation set is &lt;STRONG&gt;&lt;U&gt;0.9791&lt;/U&gt; using LeastSquares directly from the FitModel and &lt;U&gt;-1.24e+12&lt;/U&gt; using Least Squares through Generalized Regression&lt;/STRONG&gt;.&amp;nbsp;&lt;/P&gt;&lt;P&gt;&lt;STRONG&gt;I do not understand the differences in the RSquare values for the validation set.&lt;/STRONG&gt; &lt;STRONG&gt;Shouldn't the values be the same (even if there is overfitting)?&lt;/STRONG&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;3. I do think I have reason to include the terms in my model. First, my response was calculated with different data based on the different number of replications (one categorical IV) and different "formula/data" based on the different number of clusters (second categorical IV). The known theoretical formula to calculate the response takes into consideration the arrival rate (continuous IV), because that value affects the data generated.&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;I am not sure how would I do bootstrapping in JMP. But I did try different models manually (with and without the terms) and I also tried using Forward Stepwise. In both cases, the results ended up showing the two categorical IV + their interaction + the continuous IV was the best model.&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;4. Your last point is actually something that I was hoping someone would bring to this discussion.&amp;nbsp;&lt;/P&gt;&lt;P&gt;My response is not actually Normal (it is LogNormal, in fact).&amp;nbsp;&lt;/P&gt;&lt;P&gt;So, I am going to adjust that when I try Ridge or ElasticNet.&amp;nbsp;&lt;/P&gt;&lt;P&gt;&lt;STRONG&gt;But even if my response is not Normal, shouldn't the RSquare values for the Validation sets in both images I uploaded match?&lt;/STRONG&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Thank you so much for your insights!&lt;/P&gt;</description>
      <pubDate>Mon, 23 Nov 2020 22:33:53 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/Difference-in-results-between-Generalized-RSquare-and-RSquare/m-p/336741#M58419</guid>
      <dc:creator>AnnaPaula</dc:creator>
      <dc:date>2020-11-23T22:33:53Z</dc:date>
    </item>
    <item>
      <title>Re: Difference in results between Generalized RSquare and RSquare</title>
      <link>https://community.jmp.com/t5/Discussions/Difference-in-results-between-Generalized-RSquare-and-RSquare/m-p/337137#M58459</link>
      <description>&lt;P&gt;I noticed in your original residual plot that the outliers all seem to be in the validation hold-out set.&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;Generalized Regression uses a different parameterization than Fit Least Squares. See this page of &lt;A href="https://www.jmp.com/support/help/en/15.2/?os=mac&amp;amp;source=application&amp;amp;utm_source=helpmenu&amp;amp;utm_medium=application#page/jmp/nominal-factors.shtml#" target="_self"&gt;statistical details&lt;/A&gt;. It also uses a different estimation method (MLE vs OLS).&lt;/P&gt;</description>
      <pubDate>Tue, 24 Nov 2020 17:58:05 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/Difference-in-results-between-Generalized-RSquare-and-RSquare/m-p/337137#M58459</guid>
      <dc:creator>Mark_Bailey</dc:creator>
      <dc:date>2020-11-24T17:58:05Z</dc:date>
    </item>
    <item>
      <title>Re: Difference in results between Generalized RSquare and RSquare</title>
      <link>https://community.jmp.com/t5/Discussions/Difference-in-results-between-Generalized-RSquare-and-RSquare/m-p/337162#M58462</link>
      <description>&lt;P&gt;Hi&amp;nbsp;&lt;a href="https://community.jmp.com/t5/user/viewprofilepage/user-id/21986"&gt;@AnnaPaula&lt;/a&gt;&amp;nbsp;,&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp; I'm not sure I quite follow how you used the calculated seeds to generate stratification of the validation column. But, if you use the validation column platform in JMP Pro (which you have if you can access GenReg), then you can just stratify on the response (Y variable) you're fitting. If all your outliers are in the validation set, you might want to re-think how it generated the two sets.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp; Another point on this you might consider is splitting off entirely -- into another data table a test data set that is not used to fit the data. If you run multiple models on the training/validation set, you can then use the test data set to see which model is really the best at prediction, you can do this using the Model Comparison platform. Presumably, your goal is to have the best predictive capability, and you need to test that somehow. It looks like you have a large enough set you could do something like that. For example, make a validation column with, e.g. 60% train, 20% validation, 20% test. Subset the test data set into an entirely different table and then use the remaining train/validate subset to build models.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp; For your other questions about R^2 and it being negative or not quite matching up, I'd refer to what Mark said.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp; If your data is log-normal distributed, you'll want to select that option in the model specification, as that changes some of the underlying processes behind how it does the fitting.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp; When you have created a model in SLS or GenReg or PLS, you can bootstrap the Estimates by right clicking the Estimate column and selecting bootstrap. You'll want to run several thousand and then look at the distributions to see if the original Estimate for the coefficient of that term is close to a global mean of many estimations. JMP will run some calculations with slightly different starting points and therefore generate several different estimates, you can then see that distribution and determine if overall 1) the estimate from the first try was accurate, and 2) if the coefficient for the effect is really contributing a lot or not. For example, in the SLS platform, the effects are given a FDR LogWorth value to estimate the false discovery rate for that effect. If the value is &amp;gt;2 (the blue line), then there's significant evidence that the effect really does contribute to the model. On the other hand, it could be smaller, or near the value 2. If you bootstrap those FDR values, you can find the mean and range to determine if the effect is meaningful or not. I've had times where an effect looked like it could be borderline, and after bootstrapping the FDR, most of the time is was actually not crossing above 2 and was therefore not really a globally important factor in the model.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp; As a last note, I highly recommend running multiple different kinds of modeling platforms: boosted tree, bootstrap forest, NN, etc. to see if another platform might work best for your data/situation.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Hope this helps!,&lt;/P&gt;&lt;P&gt;DS&lt;/P&gt;</description>
      <pubDate>Tue, 24 Nov 2020 18:47:13 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/Difference-in-results-between-Generalized-RSquare-and-RSquare/m-p/337162#M58462</guid>
      <dc:creator>SDF1</dc:creator>
      <dc:date>2020-11-24T18:47:13Z</dc:date>
    </item>
    <item>
      <title>Re: Difference in results between Generalized RSquare and RSquare</title>
      <link>https://community.jmp.com/t5/Discussions/Difference-in-results-between-Generalized-RSquare-and-RSquare/m-p/337195#M58465</link>
      <description>&lt;P&gt;Hi&amp;nbsp;&lt;a href="https://community.jmp.com/t5/user/viewprofilepage/user-id/5358"&gt;@Mark_Bailey&lt;/a&gt;&amp;nbsp;, thank you so much for getting back to me again. I really appreciate all your help and patience with my questions.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Yes, for this specific response all my outliers are in the validation set. But for other responses I am investigating, this is not true (the outliers fall in both the validation and training sets).&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;In terms of what I am modeling: I had run a few experiments using discrete event simulation (with different seeds so that I would have different data). Based on that I calculated my four responses of interest (here I showed only one).&lt;/P&gt;&lt;P&gt;Now, I am trying to fit a regression model to those responses, because my next step will be to use an optimization model to find the best setting for my responses of interest.&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;As far as I know, the&amp;nbsp;parameterization/coding&amp;nbsp;should not impact the model statistics. Am I missing something here? (&lt;A href="http://faculty.cas.usf.edu/mbrannick/regression/anova1.html" target="_self"&gt;Coding&lt;/A&gt;)&lt;/P&gt;&lt;P&gt;Also, for linear regression and when residuals are normally distributed, OLS and MLE lead to the same coefficients. Don't they? (&lt;A href="https://openclassrooms.com/en/courses/5873596-design-effective-statistical-models-to-understand-your-data/6233001-appreciate-ordinary-least-square-and-maximum-likelihood-estimation#:~:text=The%20ordinary%20least%20square%20(OLS)%20method%20is%20tailored%20to%20the,always%20give%20a%20decent%20result.&amp;amp;text=The%20maximum%20likelihood%20estimation%20(MLE)%20method%20is%20a%20more%20general,limited%20to%20linear%20regression%20models." target="_self"&gt;OLS and MLE&lt;/A&gt;)&lt;/P&gt;&lt;P&gt;According to this &lt;A href="https://community.jmp.com/t5/Tutorials/Using-Generalized-Regression-in-JMP-Pro-to-Create-Robust-Linear/ta-p/311091" target="_self"&gt;Generalized Regression tutorial&lt;/A&gt; posted by&amp;nbsp;&lt;a href="https://community.jmp.com/t5/user/viewprofilepage/user-id/2313"&gt;@gail_massari&lt;/a&gt;&amp;nbsp;, around time 8:35, my understanding is that&amp;nbsp;&lt;a href="https://community.jmp.com/t5/user/viewprofilepage/user-id/3552"&gt;@brady_brady&lt;/a&gt;&amp;nbsp;also says that the results between the MLE and the OLS should match.&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;Finally, would you mind pointing me to the JMP help page that indicates the Generalized Regression uses MLE instead of OLS when the distribution is normal?&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;Because of the image below (taken from JMP), I was under the wrong impression that the Generalized Regression would still use OLS if the distribution chosen was Normal:&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;&lt;span class="lia-inline-image-display-wrapper lia-image-align-inline" image-alt="image.png" style="width: 427px;"&gt;&lt;img src="https://community.jmp.com/t5/image/serverpage/image-id/28587i524BDB1823CDECDF/image-dimensions/427x239?v=v2" width="427" height="239" role="button" title="image.png" alt="image.png" /&gt;&lt;/span&gt;&lt;span class="lia-inline-image-display-wrapper lia-image-align-inline" image-alt="image.png" style="width: 441px;"&gt;&lt;img src="https://community.jmp.com/t5/image/serverpage/image-id/28588i629058D29B121093/image-dimensions/441x224?v=v2" width="441" height="224" role="button" title="image.png" alt="image.png" /&gt;&lt;/span&gt;&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;Thank you!&lt;/SPAN&gt;&lt;/P&gt;</description>
      <pubDate>Tue, 24 Nov 2020 19:30:09 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/Difference-in-results-between-Generalized-RSquare-and-RSquare/m-p/337195#M58465</guid>
      <dc:creator>AnnaPaula</dc:creator>
      <dc:date>2020-11-24T19:30:09Z</dc:date>
    </item>
    <item>
      <title>Re: Difference in results between Generalized RSquare and RSquare</title>
      <link>https://community.jmp.com/t5/Discussions/Difference-in-results-between-Generalized-RSquare-and-RSquare/m-p/337200#M58466</link>
      <description>&lt;P&gt;Thank you!&lt;/P&gt;&lt;P&gt;I had no idea we were able to bootstrap the estimated in JMP. Thank you for teaching me that!&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Thank you also for all the other insights. Very helpful!&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Tue, 24 Nov 2020 19:57:58 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/Difference-in-results-between-Generalized-RSquare-and-RSquare/m-p/337200#M58466</guid>
      <dc:creator>AnnaPaula</dc:creator>
      <dc:date>2020-11-24T19:57:58Z</dc:date>
    </item>
    <item>
      <title>Re: Difference in results between Generalized RSquare and RSquare</title>
      <link>https://community.jmp.com/t5/Discussions/Difference-in-results-between-Generalized-RSquare-and-RSquare/m-p/337327#M58481</link>
      <description>&lt;P&gt;You are correct, the metrics for the whole model should not change with different parameterizations. Of course, the metrics for individual terms or parameters will change.&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;So the coefficients are not the same with different parameterizations, so their estimates must be different, too.&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;The OLS solution and the MLE optimization should lead to the same estimates under the same parameterization with a normally distributed response.&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;I misled you: the standard least squares result in GenReg is OLS. See &lt;A href="https://www.jmp.com/support/help/en/15.2/?os=mac&amp;amp;source=application&amp;amp;utm_source=helpmenu&amp;amp;utm_medium=application#page/jmp/overview-of-the-generalized-regression-personality.shtml" target="_self"&gt;this page&lt;/A&gt; for more details.&lt;/P&gt;</description>
      <pubDate>Wed, 25 Nov 2020 11:56:26 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/Difference-in-results-between-Generalized-RSquare-and-RSquare/m-p/337327#M58481</guid>
      <dc:creator>Mark_Bailey</dc:creator>
      <dc:date>2020-11-25T11:56:26Z</dc:date>
    </item>
    <item>
      <title>Re: Difference in results between Generalized RSquare and RSquare</title>
      <link>https://community.jmp.com/t5/Discussions/Difference-in-results-between-Generalized-RSquare-and-RSquare/m-p/337391#M58491</link>
      <description>&lt;P&gt;Thank you for all the details and for pointing me to more information about the GenReg.&lt;/P&gt;&lt;P&gt;But I am still confused in this case (assuming normal distribution for the response), what would explain the differences in the Generalized Rsquare values between the validation sets among the two "platforms" (Fit Least Squares and GenReg)?&lt;/P&gt;&lt;P&gt;Sorry if I misunderstood something you said.&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Thank you!&lt;/P&gt;</description>
      <pubDate>Wed, 25 Nov 2020 15:33:21 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/Difference-in-results-between-Generalized-RSquare-and-RSquare/m-p/337391#M58491</guid>
      <dc:creator>AnnaPaula</dc:creator>
      <dc:date>2020-11-25T15:33:21Z</dc:date>
    </item>
    <item>
      <title>Re: Difference in results between Generalized RSquare and RSquare</title>
      <link>https://community.jmp.com/t5/Discussions/Difference-in-results-between-Generalized-RSquare-and-RSquare/m-p/337528#M58505</link>
      <description>&lt;P&gt;I cannot say from this discussion what might cause a difference. I &amp;nbsp;suggest that you contact JMP Technical Support (&lt;A href="mailto:support@jmp.com" target="_blank"&gt;support@jmp.com&lt;/A&gt;) about the matter.&lt;/P&gt;</description>
      <pubDate>Wed, 25 Nov 2020 18:44:48 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/Difference-in-results-between-Generalized-RSquare-and-RSquare/m-p/337528#M58505</guid>
      <dc:creator>Mark_Bailey</dc:creator>
      <dc:date>2020-11-25T18:44:48Z</dc:date>
    </item>
    <item>
      <title>Re: Difference in results between Generalized RSquare and RSquare</title>
      <link>https://community.jmp.com/t5/Discussions/Difference-in-results-between-Generalized-RSquare-and-RSquare/m-p/338333#M58615</link>
      <description>Thank you, Mark! I just opened a ticket about it.&lt;BR /&gt;I will post the answer here when I have it.</description>
      <pubDate>Mon, 30 Nov 2020 18:06:09 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/Difference-in-results-between-Generalized-RSquare-and-RSquare/m-p/338333#M58615</guid>
      <dc:creator>AnnaPaula</dc:creator>
      <dc:date>2020-11-30T18:06:09Z</dc:date>
    </item>
  </channel>
</rss>

