cancel
Showing results for 
Show  only  | Search instead for 
Did you mean: 
Choose Language Hide Translation Bar
KIrishGH
Level I

Finding Significant differences

I recently harvested a crop trial and when analyzing the data I found no significant differences in the treatments in yield or RFV (relative feed value) comparing them by individual cuts as well all cumulative cuts using ANOVA. An individual found significant differences selecting only cuts 1, 3, and 5. they grouped residual variance of treatments 1,3, and 5; named that group 2 as well as grouped the residual variance of treatment 6; named that group 3. 

 

 I am wondering #1 how to group variances in jmp and #2 if this is an ethical thing to do as I feel it is manipulating the data in some way?

 

I use jmp17

2 REPLIES 2

Re: Finding Significant differences

If I'm understanding your question and your data correctly, I think you'd best be served with a Mixed Model analysis. Here's a short video for explanation.

Victor_G
Super User

Re: Finding Significant differences

Hi @KIrishGH,

 

Welcome in the Community !

 

If I understand well the question, you are interested into knowing if you can find significant statistical differences between Treatments. You can do this testing easily with the platform "Fit Y by X", by specifying the Treatments as your X and RFV as your Y variable.

 

To know which statistical test to use, there are assumptions that needs to be verified :

  • Normality assumptions of the response variable (using the distribution platform and visualizing Normal Quantile plot for example),
  • Independence of observations/results (this can be assessed with domain expertise, and using control chart on the data may reveal trends/shifts if data are not independent),
  • Equal variance between groups (this can be checked with the option in the red triangle "Unequal Variance" in the Fit Y by X platform).

In your case, you have a slight deviation from normal distribution, and other assumptions seem to be respected, so you can use a parametric test assuming equal variance and properly adjusted for multiple comparison (to avoid Type I error), like Tukey-Kramer test :

Victor_G_0-1701852121076.png

Indeed, no treatment seems to be statistically significant from the others.

You will have the same conclusion with a non-parametric test adjusted for multiple comparisons like All Pairs Steel Dwass test.

 

To group some Treatments together, you can use the "Recode" option by selecting the column Treatment, right-click and "Recode". Then you'll be able to regroup some treatments together, and perform the statistical testing with the new groups.

Just for information I did it with the informations you provided (by hiding and excluding cuts 4 and 5 and grouping Treatments 1, 3 and 5 in Group 2, Treatment 6 is Group 3 and the rest is unchanged), and I can't see any differences in the outcomes of the tests :

Victor_G_1-1701852981150.pngVictor_G_2-1701853021993.png

Datatable is attached with the scripts used.

 

On the theoritical side, unless you have strong domain expertise explanation/justification to exclude some cuts (problems in the method, abnormal results, ...) and group some Treatments together (same chemical base for some Treatments, or same way of application, ...), I wouldn't try to tweak the results until I found something statistically significant. 

 

"If you torture the data long enough, it will confess to anything" (Ronald H. Coase)

 

The situation you present does look a lot like p-hacking, tweaking and readjusting the factors just for the sake of a significant p-value. This is not a safe and sound statistical practice and doesn't make sense, as statistical significance may be important, but effect size should also be considered to understand the practical significance of the findings.
Hope this answer will help you, 

Victor GUILLER
Scientific Expertise Engineer
L'Oréal - Data & Analytics