<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Re: CROSS VALIDATION - VALIDATION COLUMN METHOD in Discussions</title>
    <link>https://community.jmp.com/t5/Discussions/CROSS-VALIDATION-VALIDATION-COLUMN-METHOD/m-p/588349#M79332</link>
    <description>&lt;P&gt;Hi&amp;nbsp;&lt;a href="https://community.jmp.com/t5/user/viewprofilepage/user-id/39387"&gt;@sreekumarp&lt;/a&gt;,&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Interesting question, and I'm afraid I won't have a definitive response regarding your question, as it depends on the dataset, types of model to consider, and practices/habits of the analyst (or person doing the analysis).&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;First, it's important to know what are the use and needs between each sets :&lt;/P&gt;&lt;UL&gt;&lt;LI&gt;Training set : Used for&amp;nbsp;&lt;SPAN&gt;the actual training of the model(s),&lt;/SPAN&gt;&lt;/LI&gt;&lt;LI&gt;Validation set : Used for model optimization (hyperparameter fine-tuning, features/threshold selection, ... for example) and model selection,&lt;/LI&gt;&lt;LI&gt;Test set : Used for generalization and predictive performance assessment of the selected model on new/unseen data.&lt;/LI&gt;&lt;/UL&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;There are several choices/methods to split your data depending on your objectives and the size of your dataset :&lt;/P&gt;&lt;UL&gt;&lt;LI&gt;&lt;A href="https://towardsdatascience.com/how-to-split-data-into-three-sets-train-validation-and-test-and-why-e50d22d3e54c" target="_self"&gt;Train/Validation/test sets&lt;/A&gt;: Fixed sets to train, optimize and assess model performances. Recommended for larger datasets.&lt;/LI&gt;&lt;LI&gt;&lt;A href="https://machinelearningmastery.com/k-fold-cross-validation/" target="_self"&gt;K-folds crossvalidation&lt;/A&gt; : Split the dataset in K folds. The model is trained K-times, and each fold is used K-1 times for training, and 1 time for validation. It enables to assess model robustness, as performances should be equivalent across all folds.&lt;/LI&gt;&lt;LI&gt;&lt;A href="https://link.springer.com/referenceworkentry/10.1007/978-0-387-30164-8_469#:~:text=Definition,a%20single%2Ditem%20test%20set" target="_self"&gt;Leave-One-Out crossvalidation&lt;/A&gt; : Extreme case of the K-fold crossvalidation, where K = N (number of observations). It is used when you have small dataset, and want to assess if your model is robust.&lt;/LI&gt;&lt;LI&gt;Autovalidation/Self Validating Ensemble Model : Instead of separating some observations in different sets, you associate each observation with a weight for training and validation (a bigger weight in training induce a lower weight in validation, meaning that this observation will be used mainly for training and less for validation), and then repeat this procedure by varying the weight. It is used for very small dataset, and/or dataset where you can't independently split some observations between different sets : for example in Design of Experiments, the set of experiments to do can be based on a model, and if so, you can't split independantly some runs between training and validation, as it will bias the model in a negative way; the runs needed for estimating parameters won't be available, hence reducing dramatically the performance of the model.&lt;/LI&gt;&lt;/UL&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;All these approaches are supported by JMP :&amp;nbsp;&lt;A href="https://www.jmp.com/support/help/en/17.1/index.shtml#page/jmp/launch-the-make-validation-column-platform.shtml" target="_blank" rel="noopener"&gt;Launch the Make Validation Column Platform (jmp.com)&lt;/A&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;As a rule of thumb, a ratio 70/20/10 is often used.&amp;nbsp;You can read the paper "&lt;A href="https://arxiv.org/pdf/2202.03326.pdf" target="_self"&gt;Optimal Ratio for Data Splitting&lt;/A&gt;" here to have more details.&amp;nbsp;Generally, the higher the number of parameters in the model, the bigger your training dataset will be, as you'll need more data to estimate precisely each of the parameters in the model, so the complexity/type of model is also something to consider when creating training/validation/test sets.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;If you have a more precise use case, maybe this could be more helpful and less general to provide you some guidance ?&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;I also highly recommend the playlist "Making Friends with Machine Learning" from Cassie Kozyrkov to learn more about models training, validation and testing :&amp;nbsp;&lt;A href="https://www.youtube.com/playlist?list=PLRKtJ4IpxJpDxl0NTvNYQWKCYzHNuy2xG" target="_blank" rel="noopener"&gt;Making Friends with Machine Learning - YouTube&lt;/A&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Hope this first answer will help you,&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
    <pubDate>Tue, 10 Jan 2023 13:28:41 GMT</pubDate>
    <dc:creator>Victor_G</dc:creator>
    <dc:date>2023-01-10T13:28:41Z</dc:date>
    <item>
      <title>CROSS VALIDATION - VALIDATION COLUMN METHOD</title>
      <link>https://community.jmp.com/t5/Discussions/CROSS-VALIDATION-VALIDATION-COLUMN-METHOD/m-p/588298#M79328</link>
      <description>&lt;P&gt;When using the validation column method for cross validation , we split the data set into training , validation and test sets. This split ratio is specified by the user. Is there any guideline /reference to decide on the split ratio (such as 60:20:20 / 70:15:15 / 50:25:25 / 80 :10:10). Is it chosen also based on the total number of observations -N ?&lt;/P&gt;</description>
      <pubDate>Thu, 08 Jun 2023 16:43:59 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/CROSS-VALIDATION-VALIDATION-COLUMN-METHOD/m-p/588298#M79328</guid>
      <dc:creator>sreekumarp</dc:creator>
      <dc:date>2023-06-08T16:43:59Z</dc:date>
    </item>
    <item>
      <title>Re: CROSS VALIDATION - VALIDATION COLUMN METHOD</title>
      <link>https://community.jmp.com/t5/Discussions/CROSS-VALIDATION-VALIDATION-COLUMN-METHOD/m-p/588349#M79332</link>
      <description>&lt;P&gt;Hi&amp;nbsp;&lt;a href="https://community.jmp.com/t5/user/viewprofilepage/user-id/39387"&gt;@sreekumarp&lt;/a&gt;,&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Interesting question, and I'm afraid I won't have a definitive response regarding your question, as it depends on the dataset, types of model to consider, and practices/habits of the analyst (or person doing the analysis).&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;First, it's important to know what are the use and needs between each sets :&lt;/P&gt;&lt;UL&gt;&lt;LI&gt;Training set : Used for&amp;nbsp;&lt;SPAN&gt;the actual training of the model(s),&lt;/SPAN&gt;&lt;/LI&gt;&lt;LI&gt;Validation set : Used for model optimization (hyperparameter fine-tuning, features/threshold selection, ... for example) and model selection,&lt;/LI&gt;&lt;LI&gt;Test set : Used for generalization and predictive performance assessment of the selected model on new/unseen data.&lt;/LI&gt;&lt;/UL&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;There are several choices/methods to split your data depending on your objectives and the size of your dataset :&lt;/P&gt;&lt;UL&gt;&lt;LI&gt;&lt;A href="https://towardsdatascience.com/how-to-split-data-into-three-sets-train-validation-and-test-and-why-e50d22d3e54c" target="_self"&gt;Train/Validation/test sets&lt;/A&gt;: Fixed sets to train, optimize and assess model performances. Recommended for larger datasets.&lt;/LI&gt;&lt;LI&gt;&lt;A href="https://machinelearningmastery.com/k-fold-cross-validation/" target="_self"&gt;K-folds crossvalidation&lt;/A&gt; : Split the dataset in K folds. The model is trained K-times, and each fold is used K-1 times for training, and 1 time for validation. It enables to assess model robustness, as performances should be equivalent across all folds.&lt;/LI&gt;&lt;LI&gt;&lt;A href="https://link.springer.com/referenceworkentry/10.1007/978-0-387-30164-8_469#:~:text=Definition,a%20single%2Ditem%20test%20set" target="_self"&gt;Leave-One-Out crossvalidation&lt;/A&gt; : Extreme case of the K-fold crossvalidation, where K = N (number of observations). It is used when you have small dataset, and want to assess if your model is robust.&lt;/LI&gt;&lt;LI&gt;Autovalidation/Self Validating Ensemble Model : Instead of separating some observations in different sets, you associate each observation with a weight for training and validation (a bigger weight in training induce a lower weight in validation, meaning that this observation will be used mainly for training and less for validation), and then repeat this procedure by varying the weight. It is used for very small dataset, and/or dataset where you can't independently split some observations between different sets : for example in Design of Experiments, the set of experiments to do can be based on a model, and if so, you can't split independantly some runs between training and validation, as it will bias the model in a negative way; the runs needed for estimating parameters won't be available, hence reducing dramatically the performance of the model.&lt;/LI&gt;&lt;/UL&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;All these approaches are supported by JMP :&amp;nbsp;&lt;A href="https://www.jmp.com/support/help/en/17.1/index.shtml#page/jmp/launch-the-make-validation-column-platform.shtml" target="_blank" rel="noopener"&gt;Launch the Make Validation Column Platform (jmp.com)&lt;/A&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;As a rule of thumb, a ratio 70/20/10 is often used.&amp;nbsp;You can read the paper "&lt;A href="https://arxiv.org/pdf/2202.03326.pdf" target="_self"&gt;Optimal Ratio for Data Splitting&lt;/A&gt;" here to have more details.&amp;nbsp;Generally, the higher the number of parameters in the model, the bigger your training dataset will be, as you'll need more data to estimate precisely each of the parameters in the model, so the complexity/type of model is also something to consider when creating training/validation/test sets.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;If you have a more precise use case, maybe this could be more helpful and less general to provide you some guidance ?&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;I also highly recommend the playlist "Making Friends with Machine Learning" from Cassie Kozyrkov to learn more about models training, validation and testing :&amp;nbsp;&lt;A href="https://www.youtube.com/playlist?list=PLRKtJ4IpxJpDxl0NTvNYQWKCYzHNuy2xG" target="_blank" rel="noopener"&gt;Making Friends with Machine Learning - YouTube&lt;/A&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Hope this first answer will help you,&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Tue, 10 Jan 2023 13:28:41 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/CROSS-VALIDATION-VALIDATION-COLUMN-METHOD/m-p/588349#M79332</guid>
      <dc:creator>Victor_G</dc:creator>
      <dc:date>2023-01-10T13:28:41Z</dc:date>
    </item>
    <item>
      <title>Re: CROSS VALIDATION - VALIDATION COLUMN METHOD</title>
      <link>https://community.jmp.com/t5/Discussions/CROSS-VALIDATION-VALIDATION-COLUMN-METHOD/m-p/588453#M79341</link>
      <description>&lt;P&gt;Hi&amp;nbsp;&lt;a href="https://community.jmp.com/t5/user/viewprofilepage/user-id/39387"&gt;@sreekumarp&lt;/a&gt;&amp;nbsp;,&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp; In addition to what&amp;nbsp;&lt;a href="https://community.jmp.com/t5/user/viewprofilepage/user-id/11568"&gt;@Victor_G&lt;/a&gt;&amp;nbsp;wrote, I would also highly recommend that you split off your Test data set and make a new data table with it. That way when you train and validate your models, you can compare them on the test holdout data table to see which model performs the best. This reduces any chance that the test data set could accidentally be used in the training or validation sets.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp; There are also other ways that you can use simulated data to train your models and then test the models on the real data. I sometimes use this approach when the original data set is small and I need to keep the correlation structure of the inputs.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Good luck!,&lt;/P&gt;&lt;P&gt;DS&lt;/P&gt;</description>
      <pubDate>Tue, 10 Jan 2023 15:20:25 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/CROSS-VALIDATION-VALIDATION-COLUMN-METHOD/m-p/588453#M79341</guid>
      <dc:creator>SDF1</dc:creator>
      <dc:date>2023-01-10T15:20:25Z</dc:date>
    </item>
    <item>
      <title>Re: CROSS VALIDATION - VALIDATION COLUMN METHOD</title>
      <link>https://community.jmp.com/t5/Discussions/CROSS-VALIDATION-VALIDATION-COLUMN-METHOD/m-p/588524#M79348</link>
      <description>&lt;P&gt;Hi&amp;nbsp;&lt;a href="https://community.jmp.com/t5/user/viewprofilepage/user-id/12549"&gt;@SDF1&lt;/a&gt;,&lt;BR /&gt;&lt;BR /&gt;One minor correction to the great addition you provided : validation set is the set used for model comparison and selection, not the test set (but sometimes test and validation names are used alternatively or confusely).&lt;BR /&gt;&lt;BR /&gt;These two sets have very different purposes :&lt;/P&gt;&lt;UL&gt;&lt;LI&gt;Test set is the holdout part of data, not used &lt;STRONG&gt;before&lt;/STRONG&gt; having selected a model, in order to provide unbiased estimation of model generalization and predictive performance.&amp;nbsp;&lt;/LI&gt;&lt;LI&gt;Validation set is a portion of data used for model fine-tuning and models comparison, in order to select the best candidate model.&lt;/LI&gt;&lt;/UL&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;An explanation is given here :&amp;nbsp;&lt;A href="https://datascience.stackexchange.com/questions/43210/can-i-use-the-test-dataset-to-select-a-model" target="_blank" rel="noopener"&gt;machine learning - Can I use the test dataset to select a model? - Data Science Stack Exchange&lt;/A&gt;&lt;/P&gt;&lt;P&gt;Two explanations to this difference in use :&lt;/P&gt;&lt;OL&gt;&lt;LI&gt;Practical one : In data science competitions, you don't have access to test dataset, so in order to create and select the best performing algorithm you have to split your data in training and validation set, and "hope" to have good performances (and generalization) on the unseen test set.&lt;/LI&gt;&lt;LI&gt;Theoritical one : If you're using test set to compare models, you're actually doing data/information leakage, as you can improve your results on the test set over time by selecting the best performing algorithm on the test set (and then, perhaps continue to fine-tune it based on performance on test data or try other models...). So your test set is no longer unbiased, as the choice of the algorithm (and perhaps other actions done after the selection) will be made on this.&lt;/LI&gt;&lt;/OL&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Test set is often the last step between model creation/development and its deployment in production or publication, so the final assessment needs to be as fair and unbiased as possible.&lt;/P&gt;&lt;P&gt;Some ressources on the sets :&amp;nbsp;&lt;A href="https://mlu-explain.github.io/train-test-validation/" target="_blank" rel="noopener"&gt;Train,Test, and Validation Sets (mlu-explain.github.io)&lt;/A&gt;&lt;/P&gt;&lt;P&gt;&lt;A href="https://www.youtube.com/watch?v=-yda0fNYYG8" target="_blank" rel="noopener"&gt;MFML 071 - What's the difference between testing and validation? - YouTube&lt;/A&gt;&lt;BR /&gt;&lt;BR /&gt;I hope this may avoid any confusion in the naming and use of the sets,&lt;/P&gt;</description>
      <pubDate>Tue, 10 Jan 2023 17:15:24 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/CROSS-VALIDATION-VALIDATION-COLUMN-METHOD/m-p/588524#M79348</guid>
      <dc:creator>Victor_G</dc:creator>
      <dc:date>2023-01-10T17:15:24Z</dc:date>
    </item>
    <item>
      <title>Re: CROSS VALIDATION - VALIDATION COLUMN METHOD</title>
      <link>https://community.jmp.com/t5/Discussions/CROSS-VALIDATION-VALIDATION-COLUMN-METHOD/m-p/588574#M79354</link>
      <description>&lt;P&gt;Hi&amp;nbsp;&lt;a href="https://community.jmp.com/t5/user/viewprofilepage/user-id/11568"&gt;@Victor_G&lt;/a&gt;&amp;nbsp;,&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp; Sorry if it wasn't clear, but we are actually referring to the same thing. When I mentioned models before, I am talking about generating models using the same test/validation setup, but also across different platforms in JMP. I never build models with any one platform in mind, but generate and tune models using the different platforms (e.g. NN, BT, Bootstrap Forest, SVM, KNN, and XGBoost) and crossvalidation with the training and validation sets. Then, after I've generated the different models I compare their ability to predict on the holdout (test) data set, a data set that none of the models have seen during their training and validation steps. Sometimes the NN works best, and sometimes it's XGBoost, or some other platform.&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp; Lastly &lt;a href="https://community.jmp.com/t5/user/viewprofilepage/user-id/39387"&gt;@sreekumarp&lt;/a&gt;&amp;nbsp;, one thing I forgot to mention in my previous post, is that it really helps to stratify your crossvalidation column using your output column -- the column your trying to predict. This keeps a similar distribution structure for your training, validation, and test data sets. Sometimes this can't be done, especially with highly imbalanced data sets, but if possible, I highly recommend it. If not, the data is randomly placed in each type of data set, and this might lead to a poorly divided validation column, which can often lead to poor models.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;DS&lt;/P&gt;</description>
      <pubDate>Tue, 10 Jan 2023 18:18:33 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/CROSS-VALIDATION-VALIDATION-COLUMN-METHOD/m-p/588574#M79354</guid>
      <dc:creator>SDF1</dc:creator>
      <dc:date>2023-01-10T18:18:33Z</dc:date>
    </item>
    <item>
      <title>Re: CROSS VALIDATION - VALIDATION COLUMN METHOD</title>
      <link>https://community.jmp.com/t5/Discussions/CROSS-VALIDATION-VALIDATION-COLUMN-METHOD/m-p/588585#M79355</link>
      <description>&lt;P&gt;Good point on the stratified sampling&amp;nbsp;&lt;a href="https://community.jmp.com/t5/user/viewprofilepage/user-id/12549"&gt;@SDF1&lt;/a&gt; !&lt;/P&gt;&lt;P&gt;I'm sorry, we don't use the same naming or there is a misunderstanding.&lt;BR /&gt;&lt;BR /&gt;If you bring several models in the test set to compare them (and select one), that's an issue and not the purpose of this set, no matter the platforms or models used. Comparing models should be done on the validation set (to keep a clear and "pure" test set without any information leakage or bias), hence my previous answer with some ressources on this topic.&lt;BR /&gt;&lt;BR /&gt;See JMP Help also emphasizing on the differences between sets :"The testing set checks the model’s predictive ability &lt;STRONG&gt;after&lt;/STRONG&gt; a model has been chosen." :&amp;nbsp;&lt;A href="https://www.jmp.com/support/help/en/17.1/index.shtml#page/jmp/overview-of-the-make-validation-column-platform.shtml#ww269465" target="_blank" rel="noopener"&gt;https://www.jmp.com/support/help/en/17.1/index.shtml#page/jmp/overview-of-the-make-validation-column-platform.shtml#ww269465&lt;/A&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;This is also a reason of the importance of well defining the different sets and either fixed them or the method, to properly evaluate different models on the same validation set (or same method).&lt;/P&gt;</description>
      <pubDate>Tue, 10 Jan 2023 19:21:29 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/CROSS-VALIDATION-VALIDATION-COLUMN-METHOD/m-p/588585#M79355</guid>
      <dc:creator>Victor_G</dc:creator>
      <dc:date>2023-01-10T19:21:29Z</dc:date>
    </item>
    <item>
      <title>Re: CROSS VALIDATION - VALIDATION COLUMN METHOD</title>
      <link>https://community.jmp.com/t5/Discussions/CROSS-VALIDATION-VALIDATION-COLUMN-METHOD/m-p/588595#M79357</link>
      <description>&lt;P&gt;Hi&amp;nbsp;&lt;a href="https://community.jmp.com/t5/user/viewprofilepage/user-id/11568"&gt;@Victor_G&lt;/a&gt;&amp;nbsp;,&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp; Yeah, I think there might be some kind of misunderstanding. Perhaps I should say that there is a hold-out data set rather than a test data set. The hold-out data set is data that was never used in training, validation, or selection of a model from a given algorithm (test set). I use this hold-out set to compare the different algorithms against each other to see which algorithm performs the best. JMP's Model Comparison platform is very helpful in comparing the different algorithms against each other using a hold-out set. This allows for an unbiased and no leakage comparison of the different algorithms. It therefore becomes a sort of "test" data set in the sense that the hold-out data set is being used to see which of the different algorithms perform best on this "pure" data set.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp; This is the only way I am aware of to compare different algorithms against each other in JMP. What I mean is that this is the only way JMP can compare a neural net algorithm against an XGBoost algorithm, for example. I would never use, nor recommend, using the hold-out data set to compare different models from within a platform. I wouldn't compare the 20 different tuned models within the bootstrap forest platform against each other using the hold-out set, that would be the purpose of the test data set.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp; So, sorry if there was any misunderstanding/confusion. I'll try to refer to it as the hold-out set from now on to avoid confusion with the test set.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Thanks for the discussion!,&lt;/P&gt;&lt;P&gt;DS&lt;/P&gt;</description>
      <pubDate>Tue, 10 Jan 2023 19:30:55 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/CROSS-VALIDATION-VALIDATION-COLUMN-METHOD/m-p/588595#M79357</guid>
      <dc:creator>SDF1</dc:creator>
      <dc:date>2023-01-10T19:30:55Z</dc:date>
    </item>
    <item>
      <title>Re: CROSS VALIDATION - VALIDATION COLUMN METHOD</title>
      <link>https://community.jmp.com/t5/Discussions/CROSS-VALIDATION-VALIDATION-COLUMN-METHOD/m-p/588813#M79374</link>
      <description>&lt;P&gt;Thank you for providing a detailed input on the splitting of the data sets in machine learning. I am sure this will help in my research.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Sreekumar Punnappilly&lt;/P&gt;</description>
      <pubDate>Wed, 11 Jan 2023 10:29:02 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/CROSS-VALIDATION-VALIDATION-COLUMN-METHOD/m-p/588813#M79374</guid>
      <dc:creator>sreekumarp</dc:creator>
      <dc:date>2023-01-11T10:29:02Z</dc:date>
    </item>
    <item>
      <title>Re: CROSS VALIDATION - VALIDATION COLUMN METHOD</title>
      <link>https://community.jmp.com/t5/Discussions/CROSS-VALIDATION-VALIDATION-COLUMN-METHOD/m-p/588816#M79375</link>
      <description>&lt;P&gt;Thank you for your input on the cross validation column.&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Sreekumar Punnappilly&lt;/P&gt;</description>
      <pubDate>Wed, 11 Jan 2023 10:33:49 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/CROSS-VALIDATION-VALIDATION-COLUMN-METHOD/m-p/588816#M79375</guid>
      <dc:creator>sreekumarp</dc:creator>
      <dc:date>2023-01-11T10:33:49Z</dc:date>
    </item>
    <item>
      <title>Re: CROSS VALIDATION - VALIDATION COLUMN METHOD</title>
      <link>https://community.jmp.com/t5/Discussions/CROSS-VALIDATION-VALIDATION-COLUMN-METHOD/m-p/588830#M79378</link>
      <description>&lt;P&gt;You're welcome&amp;nbsp;&lt;a href="https://community.jmp.com/t5/user/viewprofilepage/user-id/39387"&gt;@sreekumarp&lt;/a&gt;.&lt;BR /&gt;&lt;BR /&gt;If you consider one or several of these answers as solution(s), don't hesitate to mark them as solution(s), to help visitors of the JMP Community to more easily find the answers they are looking for.&lt;BR /&gt;If you have more questions or a concrete case on which you would like some advice, don't hesitate to answer on this topic or create a new one.&lt;BR /&gt;&lt;BR /&gt;&lt;/P&gt;</description>
      <pubDate>Wed, 11 Jan 2023 12:11:28 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/CROSS-VALIDATION-VALIDATION-COLUMN-METHOD/m-p/588830#M79378</guid>
      <dc:creator>Victor_G</dc:creator>
      <dc:date>2023-01-11T12:11:28Z</dc:date>
    </item>
    <item>
      <title>Re: CROSS VALIDATION - VALIDATION COLUMN METHOD</title>
      <link>https://community.jmp.com/t5/Discussions/CROSS-VALIDATION-VALIDATION-COLUMN-METHOD/m-p/588874#M79381</link>
      <description>&lt;P&gt;To your original question, no, there are not specific rules about how much data to leave out. In the JMP Education analytics courses, we advise you to hold out as much data as you are comfortable with, with at least 20% held out. If you feel the training set is too small to hold back that many rows, consider k-fold cross validation. How many rows are you willing to sacrifice to validation? Use k = n / that many rows. If k &amp;lt; 5 using that formula, consider leave-one-out cross validation.&lt;/P&gt;</description>
      <pubDate>Wed, 11 Jan 2023 13:31:22 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/CROSS-VALIDATION-VALIDATION-COLUMN-METHOD/m-p/588874#M79381</guid>
      <dc:creator>Di_Michelson</dc:creator>
      <dc:date>2023-01-11T13:31:22Z</dc:date>
    </item>
    <item>
      <title>Re: CROSS VALIDATION - VALIDATION COLUMN METHOD</title>
      <link>https://community.jmp.com/t5/Discussions/CROSS-VALIDATION-VALIDATION-COLUMN-METHOD/m-p/937143#M109242</link>
      <description>&lt;P&gt;Thanks.&amp;nbsp;&lt;BR /&gt;&lt;BR /&gt;There are actually two points to be clarified for K-fold cross-validation (CV) and leave-one-out (LOO) validation using routine regression in JMP Pro.&amp;nbsp;&lt;BR /&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;Both methods are directly available in the PLSR part (partial least-squares regression), but there's a certain unclarity what to do and how to visualize the metrics if CV and/or LOO are applied for a simple regression. For example, I'm analyzing spectra and PCR and PLSR are routinely involved in this chemometric process.&lt;BR /&gt;&lt;span class="lia-inline-image-display-wrapper lia-image-align-inline" image-alt="4-fold validation.png" style="width: 729px;"&gt;&lt;img src="https://community.jmp.com/t5/image/serverpage/image-id/96968i384A80B0C82A7ECB/image-dimensions/729x256?v=v2" width="729" height="256" role="button" title="4-fold validation.png" alt="4-fold validation.png" /&gt;&lt;/span&gt;&lt;BR /&gt;I've got a little strange experience.&amp;nbsp;&amp;nbsp;However, after I create a 4-fold validation column (above, 4 folds across 72 observations) and run PCR (numeric response and principal components as predictors, I receive the following error response):&lt;/P&gt;
&lt;P&gt;&lt;BR /&gt;&lt;span class="lia-inline-image-display-wrapper lia-image-align-inline" image-alt="4-fold validation1.png" style="width: 826px;"&gt;&lt;img src="https://community.jmp.com/t5/image/serverpage/image-id/96969i9ACAB8291B12FC76/image-dimensions/826x224?v=v2" width="826" height="224" role="button" title="4-fold validation1.png" alt="4-fold validation1.png" /&gt;&lt;/span&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;Hence, there's no real CV performed, right?&amp;nbsp;&lt;BR /&gt;&lt;BR /&gt;The second point: how to build a LOO validation (a column?) for PCR?&lt;BR /&gt;&lt;BR /&gt;As the whole, please, advise me what has to be done to perform adequately cross-validation and leave-one-out validation for a typical regression and get proper metrics of fit? It's really important and unobvious yet to me.&amp;nbsp;&lt;BR /&gt;Thanks a lot, colleagues!&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Tue, 24 Mar 2026 15:09:53 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/CROSS-VALIDATION-VALIDATION-COLUMN-METHOD/m-p/937143#M109242</guid>
      <dc:creator>Nazarkovsky</dc:creator>
      <dc:date>2026-03-24T15:09:53Z</dc:date>
    </item>
    <item>
      <title>Re: CROSS VALIDATION - VALIDATION COLUMN METHOD</title>
      <link>https://community.jmp.com/t5/Discussions/CROSS-VALIDATION-VALIDATION-COLUMN-METHOD/m-p/937275#M109261</link>
      <description>&lt;P&gt;Hi&amp;nbsp;&lt;a href="https://community.jmp.com/t5/user/viewprofilepage/user-id/19209"&gt;@Nazarkovsky&lt;/a&gt;,&lt;/P&gt;
&lt;P&gt;It would have been probably easier to start a new post instead of replying to this old post, as it would have provide more visibility and enable other JMP users to join the discussion about your question.&lt;/P&gt;
&lt;P&gt;Yes, there are sometimes some errors in JMP modeling platforms when using fixed K-Folds crossvalidation column, as JMP is mostly recognizing 2 levels (training and validation or test) or 3 levels (training, validation and test) from validation column. In the error message you have, no CV is done, a linear model is fit on all your data.&lt;BR /&gt;Since you're using a very simple linear model, you can perhaps try to use&amp;nbsp;&lt;A href="https://www.jmp.com/support/help/en/18.1/#page/jmp/model-screening.shtml?_gl=1*s54ft*_up*MQ..*_ga*MzU2OTAwNjE5LjE3NzQ0MzEwNTg.*_ga_BRNVBEC1RS*czE3NzQ0MzEwNTckbzEkZzAkdDE3NzQ0MzEwNTckajYwJGwwJGgw#" target="_blank" rel="noopener"&gt;Model Screening&lt;/A&gt;&amp;nbsp;platform (which also includes PLS if you want to compare the different model's results), specifying your X's, your response Y, your 4-folds validation column (or doing the crossvalidation directly from the platform), and the type of model you want to fit (linear regression models options + modeling options):&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&lt;span class="lia-inline-image-display-wrapper lia-image-align-inline" image-alt="Capture d'écran 2026-03-25 103300.png" style="width: 400px;"&gt;&lt;img src="https://community.jmp.com/t5/image/serverpage/image-id/97060iD1309145A69E4320/image-size/medium?v=v2&amp;amp;px=400" role="button" title="Capture d'écran 2026-03-25 103300.png" alt="Capture d'écran 2026-03-25 103300.png" /&gt;&lt;/span&gt;&lt;/P&gt;
&lt;P&gt;Or directly with the K-folds crossvalidation option from the platform, you can obtain these results:&lt;/P&gt;
&lt;P&gt;&lt;span class="lia-inline-image-display-wrapper lia-image-align-inline" image-alt="Capture d'écran 2026-03-25 130627.png" style="width: 400px;"&gt;&lt;img src="https://community.jmp.com/t5/image/serverpage/image-id/97077iC247DDF3ACB5895C/image-size/medium?v=v2&amp;amp;px=400" role="button" title="Capture d'écran 2026-03-25 130627.png" alt="Capture d'écran 2026-03-25 130627.png" /&gt;&lt;/span&gt;&lt;/P&gt;
&lt;P&gt;If you want to run a Leave-One-Out cross-validation, specify simply K=n (n being the number of runs in your table). If you want to have access to individual folds results, you can adapt the solution I provided here to your modeling platform:&amp;nbsp;&lt;LI-MESSAGE title="Accessing out of fold metrics for K-Fold CV" uid="932392" url="https://community.jmp.com/t5/Discussions/Accessing-out-of-fold-metrics-for-K-Fold-CV/m-p/932392#U932392" discussion_style_icon_css="lia-mention-container-editor-message lia-img-icon-forum-thread lia-fa-icon lia-fa-forum lia-fa-thread lia-fa"&gt;&lt;/LI-MESSAGE&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;Finally, I see at least one major problem in your modeling workflow that could lead to data leakage : you need to apply the cross-validation or your splitting/validation strategy on your principal component analysis too. If you don't do it, the PCA will see the entire dataset and learn the correlations between the factors from all the dataset, and the linear model you're fitting after (and evaluate thanks to cross-validation) will benefit from the entire information from the dataset, not the information from 3 out of 4 folds. Since JMP hasn't implemented a validation role (yet !) for PCA, this is why I have added this Wishlist, to prevent data leakage and possible errors in modeling workflows like yours:&amp;nbsp;&lt;LI-MESSAGE title="Add validation role option in Principal Component Analysis platform" uid="638160" url="https://community.jmp.com/t5/JMP-Wish-List/Add-validation-role-option-in-Principal-Component-Analysis/m-p/638160#U638160" discussion_style_icon_css="lia-mention-container-editor-message lia-img-icon-idea-thread lia-fa-icon lia-fa-idea lia-fa-thread lia-fa"&gt;&lt;/LI-MESSAGE&gt;&amp;nbsp;To summarize, you should :&lt;/P&gt;
&lt;OL&gt;
&lt;LI&gt;Define and apply a validation strategy (CV, LOO or standard train/validation or train/validation/test split).&lt;/LI&gt;
&lt;LI&gt;Preprocess the data only on training set (or preprocess the data several time depending on the folds used)&lt;/LI&gt;
&lt;LI&gt;Fit a model using the same validation strategy with the preprocessed data from training set (or from the corresponding training folds).&lt;/LI&gt;
&lt;/OL&gt;
&lt;P&gt;Hope this answer will help you,&lt;/P&gt;</description>
      <pubDate>Wed, 25 Mar 2026 12:07:20 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/CROSS-VALIDATION-VALIDATION-COLUMN-METHOD/m-p/937275#M109261</guid>
      <dc:creator>Victor_G</dc:creator>
      <dc:date>2026-03-25T12:07:20Z</dc:date>
    </item>
    <item>
      <title>Re: CROSS VALIDATION - VALIDATION COLUMN METHOD</title>
      <link>https://community.jmp.com/t5/Discussions/CROSS-VALIDATION-VALIDATION-COLUMN-METHOD/m-p/937334#M109265</link>
      <description>&lt;P&gt;Anyway, the section for k-fold crossvalidation in JMP is still weak. I feel curious why it doesn't concern the developers, as this validation method is a primary tool for a preliminary assessment of any model, especially at low number of data. Moreover, in the proper option to build a K-fold CV column, it's impossible to build 3-fold cross-validation - the minimal value is 4.&amp;nbsp;&lt;BR /&gt;&lt;BR /&gt;I also do not like JMP's report of the metrics for CV, since there is no final option/button to build a 'Predicted' column after all the runs as an average result and compare to the predicted values delivered by Generalized Regression or PLSR.&lt;BR /&gt;&lt;BR /&gt;As for the validation of PCA, this option is encountered in Unscrambler, a traditional software for chemometrics (multivariate analysis of chemical data, like spectra). Yes, the validation is scanned across the pre-set numbers of components vs. explained variance. For this test of leakage I use Unscrambler, aha.&lt;BR /&gt;&lt;BR /&gt;There's a series of JMP posts here in Community dedicated to the analysis of spectra, but still short in comprehensive discussion for CV and PCR.&amp;nbsp;&lt;BR /&gt;&lt;BR /&gt;However, in chemometrics, SVR (support vector regression), PLSR, and PCR are always applied and their metrics are compared.&lt;BR /&gt;&lt;BR /&gt;For LOO thanks, it really works properly.&lt;/P&gt;</description>
      <pubDate>Wed, 25 Mar 2026 17:17:47 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/CROSS-VALIDATION-VALIDATION-COLUMN-METHOD/m-p/937334#M109265</guid>
      <dc:creator>Nazarkovsky</dc:creator>
      <dc:date>2026-03-25T17:17:47Z</dc:date>
    </item>
    <item>
      <title>Re: CROSS VALIDATION - VALIDATION COLUMN METHOD</title>
      <link>https://community.jmp.com/t5/Discussions/CROSS-VALIDATION-VALIDATION-COLUMN-METHOD/m-p/937335#M109266</link>
      <description>&lt;P&gt;Anyway, the part k-fold crossvalidation in JMP is still weak. I feel curious why it doesn't concern the developers, as this validation method is a primary tool for a preliminar assessing of any model, especially at low number of data. Moreover, in a proper option to build a K-fold CV column, it's impossible to build 3-fold cross-validation - the minimal value is 4.&amp;nbsp;&lt;BR /&gt;I also do not like JMP's report of the metrics for CV, since there is no final option/button to build a 'Predicted' column after all the runs as an average result and compare to the predicted values delivered by Generalized Regression or PLSR.&lt;BR /&gt;As for the validation of PCA, this option is encountered in Unscrambler, a traditional software for chemometrics (multivariate analysis of chemical data, like spectra). Yes, the validation is scanned acrosse the set numbers of components vs. explained variance. For this test of leakage I use Unscrambler, aha.&lt;BR /&gt;There's a series of JMP posts here in Community dedicated to the analysis of spectra, but still short in comprehensive discussion for CV and PCR.&amp;nbsp;&lt;BR /&gt;However, in chemometrics, SVR (support vector regression), PLSR, and PCR are always applied and their metrics are compared.&lt;BR /&gt;For LOO thanks, it really works properly.&lt;/P&gt;</description>
      <pubDate>Wed, 25 Mar 2026 15:57:48 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/CROSS-VALIDATION-VALIDATION-COLUMN-METHOD/m-p/937335#M109266</guid>
      <dc:creator>Nazarkovsky</dc:creator>
      <dc:date>2026-03-25T15:57:48Z</dc:date>
    </item>
  </channel>
</rss>

