<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Re: Testing for Normality in Discussions</title>
    <link>https://community.jmp.com/t5/Discussions/Testing-for-Normality/m-p/414856#M66391</link>
    <description>&lt;P&gt;Most of the statistics we are taught were developed 80-100 years ago, when small samples were the rule. So the issue was, "What can I convincingly learn from or decide with a small sample?" Inferential techniques started then. Such situations still exist (e.g., no data available, need to experiment...) but a new situation arose 25 years ago. We now have massive databases from which to learn. Inference is pointless. Everything is statistically significant. That is, even the smallest differences or smallest parameter estimates are statistically significant. So, predictive modeling does not try to infer significance. Instead, the focus is on predictive performance through feature / model selection. It is a different challenge and requires a different mindset.&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;Also, in this specific case, why is a normal distribution assumed? What is supposed to be normally distributed? Data? An estimate? And what are the consequences of non-normality? How much of a departure from normal is necessary to compromise the result or decision? How robust is the method?&lt;/P&gt;</description>
    <pubDate>Thu, 02 Sep 2021 14:51:18 GMT</pubDate>
    <dc:creator>Mark_Bailey</dc:creator>
    <dc:date>2021-09-02T14:51:18Z</dc:date>
    <item>
      <title>Testing for Normality</title>
      <link>https://community.jmp.com/t5/Discussions/Testing-for-Normality/m-p/414327#M66344</link>
      <description>&lt;P&gt;I wanted to obtain Shapiro Wilks Test for a distribution but I cannot obtain it. I tried the method posted on the discussion about legacy fitters. That only gives me KSL test. Otherwise if I do not use legacy fitters it gives me Anderson-Darlling test. My sample size is extremely large, &amp;gt;30k, not sure if that has anything to do with it. Thank you for help&lt;/P&gt;</description>
      <pubDate>Fri, 09 Jun 2023 00:38:15 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/Testing-for-Normality/m-p/414327#M66344</guid>
      <dc:creator>jmpquestions123</dc:creator>
      <dc:date>2023-06-09T00:38:15Z</dc:date>
    </item>
    <item>
      <title>Re: Testing for Normality</title>
      <link>https://community.jmp.com/t5/Discussions/Testing-for-Normality/m-p/414459#M66357</link>
      <description>&lt;P&gt;For the legacy distributions, the Shapiro-Wilk test for normality is reported when the sample size is less than or equal to 2000. The KSL test is computed for samples that are greater than 2000.&amp;nbsp; When the sample size is large, the Shapiro-Wilk test has a large power. Hence any small difference between your distribution and the null hypothesis is meaningful and leads to a rejection of the null hypothesis.&lt;/P&gt;</description>
      <pubDate>Wed, 01 Sep 2021 11:29:10 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/Testing-for-Normality/m-p/414459#M66357</guid>
      <dc:creator>tonya_mauldin</dc:creator>
      <dc:date>2021-09-01T11:29:10Z</dc:date>
    </item>
    <item>
      <title>Re: Testing for Normality</title>
      <link>https://community.jmp.com/t5/Discussions/Testing-for-Normality/m-p/414496#M66362</link>
      <description>&lt;P&gt;Thank you for your message. So between KSL and Anderson-Darling, which test is a good substitute for Shapiro Wilks?&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Wed, 01 Sep 2021 13:47:27 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/Testing-for-Normality/m-p/414496#M66362</guid>
      <dc:creator>jmpquestions123</dc:creator>
      <dc:date>2021-09-01T13:47:27Z</dc:date>
    </item>
    <item>
      <title>Re: Testing for Normality</title>
      <link>https://community.jmp.com/t5/Discussions/Testing-for-Normality/m-p/414497#M66363</link>
      <description>&lt;P&gt;Anderson-Darling is a more powerful test and is the one we suggest be used.&lt;/P&gt;</description>
      <pubDate>Wed, 01 Sep 2021 14:07:32 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/Testing-for-Normality/m-p/414497#M66363</guid>
      <dc:creator>tonya_mauldin</dc:creator>
      <dc:date>2021-09-01T14:07:32Z</dc:date>
    </item>
    <item>
      <title>Re: Testing for Normality</title>
      <link>https://community.jmp.com/t5/Discussions/Testing-for-Normality/m-p/414540#M66366</link>
      <description>&lt;P&gt;I am confused on how to interpret it. All my Anderson-Darling p-values for ~50 variables are significant but the distributions of many of them look normal....&lt;/P&gt;</description>
      <pubDate>Wed, 01 Sep 2021 14:43:17 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/Testing-for-Normality/m-p/414540#M66366</guid>
      <dc:creator>jmpquestions123</dc:creator>
      <dc:date>2021-09-01T14:43:17Z</dc:date>
    </item>
    <item>
      <title>Re: Testing for Normality</title>
      <link>https://community.jmp.com/t5/Discussions/Testing-for-Normality/m-p/414591#M66371</link>
      <description>&lt;P&gt;&lt;a href="https://community.jmp.com/t5/user/viewprofilepage/user-id/1450"&gt;@tonya_mauldin&lt;/a&gt;&amp;nbsp;explained this very well when the changes were first made in a &lt;A href="https://community.jmp.com/t5/JMPer-Cable/Modernizing-the-Distribution-platform-Options-for-fitted/ba-p/233180" target="_self"&gt;JMP Blog post&lt;/A&gt;.&lt;/P&gt;</description>
      <pubDate>Wed, 01 Sep 2021 15:57:08 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/Testing-for-Normality/m-p/414591#M66371</guid>
      <dc:creator>Mark_Bailey</dc:creator>
      <dc:date>2021-09-01T15:57:08Z</dc:date>
    </item>
    <item>
      <title>Re: Testing for Normality</title>
      <link>https://community.jmp.com/t5/Discussions/Testing-for-Normality/m-p/414593#M66372</link>
      <description>&lt;P&gt;Your large sample N &amp;gt; 30,000 leads to detection of non-normal features (departures from ideal normal CDF) that might be statistically significant but practically unimportant.&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Wed, 01 Sep 2021 16:00:20 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/Testing-for-Normality/m-p/414593#M66372</guid>
      <dc:creator>Mark_Bailey</dc:creator>
      <dc:date>2021-09-01T16:00:20Z</dc:date>
    </item>
    <item>
      <title>Re: Testing for Normality</title>
      <link>https://community.jmp.com/t5/Discussions/Testing-for-Normality/m-p/414766#M66381</link>
      <description>&lt;P&gt;Thank you&amp;nbsp;&lt;a href="https://community.jmp.com/t5/user/viewprofilepage/user-id/5358"&gt;@Mark_Bailey&lt;/a&gt; for this comment. This is a very important point with all statistical significance testing. unfortunately, too many people are fixed on p values without looking at visuals or the size of the coefficients. In some research fields this is partially avoided by using minimal sample sizes to detect a practically meaningful difference.&lt;/P&gt;
&lt;P&gt;Yet, it is very difficult to explain to some people, including reviewers and researchers, that it is very easy to get statistical significance that is meaningless. &amp;nbsp; &lt;/P&gt;</description>
      <pubDate>Thu, 02 Sep 2021 11:01:55 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/Testing-for-Normality/m-p/414766#M66381</guid>
      <dc:creator>ron_horne</dc:creator>
      <dc:date>2021-09-02T11:01:55Z</dc:date>
    </item>
    <item>
      <title>Re: Testing for Normality</title>
      <link>https://community.jmp.com/t5/Discussions/Testing-for-Normality/m-p/414856#M66391</link>
      <description>&lt;P&gt;Most of the statistics we are taught were developed 80-100 years ago, when small samples were the rule. So the issue was, "What can I convincingly learn from or decide with a small sample?" Inferential techniques started then. Such situations still exist (e.g., no data available, need to experiment...) but a new situation arose 25 years ago. We now have massive databases from which to learn. Inference is pointless. Everything is statistically significant. That is, even the smallest differences or smallest parameter estimates are statistically significant. So, predictive modeling does not try to infer significance. Instead, the focus is on predictive performance through feature / model selection. It is a different challenge and requires a different mindset.&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;Also, in this specific case, why is a normal distribution assumed? What is supposed to be normally distributed? Data? An estimate? And what are the consequences of non-normality? How much of a departure from normal is necessary to compromise the result or decision? How robust is the method?&lt;/P&gt;</description>
      <pubDate>Thu, 02 Sep 2021 14:51:18 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/Testing-for-Normality/m-p/414856#M66391</guid>
      <dc:creator>Mark_Bailey</dc:creator>
      <dc:date>2021-09-02T14:51:18Z</dc:date>
    </item>
  </channel>
</rss>

