cancel
Showing results for 
Show  only  | Search instead for 
Did you mean: 
  • Register to attend Discovery Summit 2025 Online: Early Users Edition, Sept. 24-25.
  • New JMP features coming to desktops everywhere this September. Sign up to learn more at jmp.com/launch.
Choose Language Hide Translation Bar
dlehman1
Level VI

Conceptual question about validation

I'm hoping that people with a better statistical background than I have can shed some light on this question.  JMP has robust validation capabilities for all predictive modeling, including cross validation (though not in the Fit Model platform, except for the capability of using a validation set).  Data scientists routinely partition their data into training and validation (and sometimes test) data sets.  However, most published work using traditional statistical methods (multiple regression, logistic regression) do not use any validation.  I've been wondering why that is the case.  Among the possibilities I can think of:  (1) statistical significance is viewed as making validation unnecessary; (2) historical practice that is slow to change; (3) the emphasis is on inference and not prediction (along the lines of Brieman's two cultures).  In any case, I don't understand any justification for not using validation, so I'd like to hear what people think.

11 REPLIES 11
P_Bartell
Level VIII

Re: Conceptual question about validation

I concur with everything that @statman and @Jed_Campbell have contributed. The only thing I'll add is maybe there is a fourth 'reason'? Until model validation is taught as a requirement of the extended problem solving process of which modeling is a component you just won't see much of it included in the literature. Just another piece or consequence of the 'historical practice that is slow to change.'

SDF1
Super User

Re: Conceptual question about validation

Hi @dlehman1 ,

 

  Thanks for posing an interesting conceptual question. In general, I agree with what everyone has commented on so far -- that it's important to define what "validation" means in this case. However, I would say that validation can take on both of those meanings -- a hold out set to improve model prediction as well as generating new samples that test the predictive capability of the model. 

 

  To me, both are important and need to be done -- one for improving the model, and the other to make sure the model can work, especially when generating new samples that are near the boundaries of the parameter space where models tend to be less accurate.

 

  As to why published works tend not to use validation is to a large extent what you and @Jed_Campbell commented on: historical practice is slow to change -- but this is more of a result of influences within the scientific community that push toward finding a significant result rather than a robust model. My background is physics, and physicists are notorious for creating simple models for one situation and then extrapolating that out to other situations. Take the simple pendulum as an example. This is the basis for almost all basic physics problems, yet in practice it isn't the greatest model -- there have been so many tweaks and changes to it in order to have the model work in other non-ideal situations (think about having to add the electron spin into the orbital mechanics of the electronic states of an atom).

 

  To me, this simple pendulum doesn't make a very robust model -- but it does satisfy the interest and desire to "find" something significant in the data. Sure, it's a great start, and we all need to start somewhere. However, if it comes to wanting to generate a broader, more robust model that can be utilized in more generalized areas, validation is required in both definitions that were discussed.

 

  More "mundane" results might ultimately lead to more robust, better models and predictive capabilities in science, but cultural changes in the scientific community need to change, and a de-emphasizing of the splashy, flashy new findings probably needs to happen. 

 

  In short, I would say that not only is historical practice slow to change, but there also needs to be a cultural change in the scientific community about what/how results should be recognized in science. A sort of reorienting of priorities.

 

DS

Recommended Articles