I'm curious if there are any times when you should not use the predictor screening tool, & if so, what should you check with your data to understand if you might have issues using predictor screening?
I find predictor screening immensely valuable, but like any automated tool it can't account for some things. It is most valuable when you have many potential predictors - however, when these predictors are closely related, I don't quite trust the automated procedure. I find this to be a common problem, where I have data sets that have a number of closely related measures (for example, hotel availability at a number of different time points). If you just use predictor screening you may get the single most closely related one of the measures or possibly a number of them will be chosen. Or, potentially none will stand out (I'm not sure about that one)? What I do in such cases is manually analyze the closely related predictors, using the Multivariate platform. I can then choose one of the related measures to use in predictor screening.
There are other more complicated cases. Predictor screening will not identify if you should create new variables as functions of existing ones (such as ratios of two measures). It will not give you a good idea whether you should recode some of your nominal variables (I'm not sure about this one either - I haven't experimented to see). In general, I'd say that Predictor Screening does not replace all of the normal data cleaning steps - it is most valuable when used after cleaning the data.
Hi @mjz5448,
I might add a complementary answer to elucidate some of the informations given, based on the previous responses I have already written on similar topics:
Predictor Screening is a method based on the Random Forest algorithm that help you identify important variables (identified by the calculation of features importance) for your response(s) in your historical data. Depending on the representativeness of your data, coverage of the experimental space, missing values, presence of outliers, interactions, correlations or multicollinearity between your factors, this platform may have some shortcuts, many unknowns, and shouldn't be trust blindly (as stated by @dlehman1).
Concerning your questions about multicollinearity :
More infos on tree-based models :
And about the advantages of Random Forests : https://www.linkedin.com/posts/victorguiller_doe-machinelearning-randomforests-activity-712755779981...
Hope this complementary answer will help you,
I find predictor screening immensely valuable, but like any automated tool it can't account for some things. It is most valuable when you have many potential predictors - however, when these predictors are closely related, I don't quite trust the automated procedure. I find this to be a common problem, where I have data sets that have a number of closely related measures (for example, hotel availability at a number of different time points). If you just use predictor screening you may get the single most closely related one of the measures or possibly a number of them will be chosen. Or, potentially none will stand out (I'm not sure about that one)? What I do in such cases is manually analyze the closely related predictors, using the Multivariate platform. I can then choose one of the related measures to use in predictor screening.
There are other more complicated cases. Predictor screening will not identify if you should create new variables as functions of existing ones (such as ratios of two measures). It will not give you a good idea whether you should recode some of your nominal variables (I'm not sure about this one either - I haven't experimented to see). In general, I'd say that Predictor Screening does not replace all of the normal data cleaning steps - it is most valuable when used after cleaning the data.
Thanks. I also find it pretty invaluable, so I'm wondering where it might fail, & how to check for that failure? I don't know much about the bootstrap forest model that (I think) it's based on, but have briefly read that it's less susceptible to multicollinearity the way linear models are, but have also ready that it could fail if you have linear relationships with the response and predictors, or if you lack iid, but I don't know how you check those assumptions in your predictor screener, or if that is in fact the case as I'm fairly new to these types of data analysis. It would be nice if JMP had a tutorial on what to look for in terms of assumption checking etc..
Hi @mjz5448,
I might add a complementary answer to elucidate some of the informations given, based on the previous responses I have already written on similar topics:
Predictor Screening is a method based on the Random Forest algorithm that help you identify important variables (identified by the calculation of features importance) for your response(s) in your historical data. Depending on the representativeness of your data, coverage of the experimental space, missing values, presence of outliers, interactions, correlations or multicollinearity between your factors, this platform may have some shortcuts, many unknowns, and shouldn't be trust blindly (as stated by @dlehman1).
Concerning your questions about multicollinearity :
More infos on tree-based models :
And about the advantages of Random Forests : https://www.linkedin.com/posts/victorguiller_doe-machinelearning-randomforests-activity-712755779981...
Hope this complementary answer will help you,
Thanks Victor_G. I appreciate the insight into multicollinearity - it's not something I typically account for, but will try to do so going forward.
Do you have any insight on whether predictor screening will work if you data is not iid (independent & identically distributed)? I've read that tree-based/random forest models have limitations if your data has a time-series trend, which can be overcome, but that you some how need to transform the data. I'm not sure if that's applicable to predictor screening or not though as I'm just learning about this topic.
Hi @mjz5448,
Random Forests can be used for time series forecasting (and other algorithms too).
The biggest precaution with non-iid data is to take into account the dependencies/groups in the data, whether it comes from groups of similar observations/individuals, time dependency, measurement repetition, etc...., in your data splitting (training/validation/test). Incorrect data splitting for non-iid data could lead to data leakage, false expectations and over-optimistic results.
Typically for time series forecasting, a specific data splitting for training and validation should be done (not randomly, but more in a "time-window" style), so that the time structure dependency of the responses/data is respected and that there is no data leakage in the setup : no data "from the future" is used in the training that would lead to over-optimistic predictions.
Some reading ressources :
https://machinelearningmastery.com/random-forest-for-time-series-forecasting/
https://mindfulmodeler.substack.com/p/how-to-deal-with-non-iid-data-in
Hope this answer will help you,
Thanks Victor_G. I will take a look.