I'm doing what is basic least square regression on 12 points (4 points in triplicate) but I get different results between excel and JMP. Can anyone offer an explanation? Thanks
|UCL (99.7%) at intercept||3.110484||10.0827...|
for data the data below:
Quick addendum: I get the same result as excel when using the JMP excel pluggin but different when putting the numbers directly into JMP!
Friends don't let friends use Excel for statistics! There are many scholarly articles available about the statistical problems with Excel. Microsoft knows about them, yet won't fix them. Let the consumer beware.
In this case, my guess is that the data in Excel has greater precision then the 3 decimal places shown in the original post.
I checked that bit and changed all the data to be strictly the same amount of decimal places
mmmhhhhh I get the same answer as well now....... something strange is going on somewhere.
Anyway! Thanks for the help!