<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Is there a way to do k-fold cross validation with boosted tree? in Discussions</title>
    <link>https://community.jmp.com/t5/Discussions/Is-there-a-way-to-do-k-fold-cross-validation-with-boosted-tree/m-p/781144#M96362</link>
    <description>&lt;P&gt;Is there a way to do k-fold cross validation with the boosted tree platform?&lt;/P&gt;&lt;P&gt;While I understand this can be done in model screening, I need to hypertune with the k-fold that provides the maximum performance. So, instead of using model screening, is there a way to do k-fold cross validation with boosted tree predictive modeling where I can also hypertune other parameters?&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Thank you for taking the time to get back to me.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Nisha&lt;/P&gt;</description>
    <pubDate>Mon, 12 Aug 2024 18:46:56 GMT</pubDate>
    <dc:creator>NishaKumar2023</dc:creator>
    <dc:date>2024-08-12T18:46:56Z</dc:date>
    <item>
      <title>Is there a way to do k-fold cross validation with boosted tree?</title>
      <link>https://community.jmp.com/t5/Discussions/Is-there-a-way-to-do-k-fold-cross-validation-with-boosted-tree/m-p/781144#M96362</link>
      <description>&lt;P&gt;Is there a way to do k-fold cross validation with the boosted tree platform?&lt;/P&gt;&lt;P&gt;While I understand this can be done in model screening, I need to hypertune with the k-fold that provides the maximum performance. So, instead of using model screening, is there a way to do k-fold cross validation with boosted tree predictive modeling where I can also hypertune other parameters?&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Thank you for taking the time to get back to me.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Nisha&lt;/P&gt;</description>
      <pubDate>Mon, 12 Aug 2024 18:46:56 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/Is-there-a-way-to-do-k-fold-cross-validation-with-boosted-tree/m-p/781144#M96362</guid>
      <dc:creator>NishaKumar2023</dc:creator>
      <dc:date>2024-08-12T18:46:56Z</dc:date>
    </item>
    <item>
      <title>Re: Is there a way to do k-fold cross validation with boosted tree?</title>
      <link>https://community.jmp.com/t5/Discussions/Is-there-a-way-to-do-k-fold-cross-validation-with-boosted-tree/m-p/781176#M96372</link>
      <description>&lt;P&gt;I'd try using the XGBoost app if you'd like to use K-fold and boosted tree.&lt;/P&gt;&lt;P&gt;&lt;A href="https://community.jmp.com/t5/JMP-Add-Ins/XGBoost-Add-In-for-JMP-Pro/ta-p/319383" target="_blank"&gt;XGBoost Add-In for JMP Pro - JMP User Community&lt;/A&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Steve&lt;/P&gt;</description>
      <pubDate>Mon, 12 Aug 2024 20:49:08 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/Is-there-a-way-to-do-k-fold-cross-validation-with-boosted-tree/m-p/781176#M96372</guid>
      <dc:creator>shampton82</dc:creator>
      <dc:date>2024-08-12T20:49:08Z</dc:date>
    </item>
    <item>
      <title>Re: Is there a way to do k-fold cross validation with boosted tree?</title>
      <link>https://community.jmp.com/t5/Discussions/Is-there-a-way-to-do-k-fold-cross-validation-with-boosted-tree/m-p/781267#M96381</link>
      <description>&lt;P&gt;Hi&amp;nbsp;&lt;a href="https://community.jmp.com/t5/user/viewprofilepage/user-id/58571"&gt;@NishaKumar2023&lt;/a&gt;,&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;You can use K-folds cross-validation while tuning your hyperparameters, but it doesn't work exactly as you may intend to do. Here is the methodology :&lt;/P&gt;
&lt;OL&gt;
&lt;LI&gt;Create a K-folds validation column (either fixed or a formula depending on your objectives and reproducibility needs) :&amp;nbsp;&lt;A href="https://www.jmp.com/support/help/en/18.0/index.shtml#page/jmp/launch-the-make-validation-column-platform.shtml#" target="_blank" rel="noopener"&gt;Launch the Make Validation Column Platform (jmp.com)&lt;/A&gt;&lt;BR /&gt;Make sure your data splitting in folds is representative and balanced, by using stratification, or that you do respect the constraint/duplication of your data by using grouping (same ID in the same fold for example).&lt;/LI&gt;
&lt;LI&gt;Open or create a Tuning datatable for the Boosted Tree (or any other platform) you would like to launch. I added in copy an example for the Boosted Tree, but you can find others tuning table provided by &lt;a href="https://community.jmp.com/t5/user/viewprofilepage/user-id/12549"&gt;@SDF1&lt;/a&gt;&amp;nbsp;in this post: &lt;LI-MESSAGE title="Malfunction in Bootstrap Forest with Tuning Design Table?" uid="715269" url="https://community.jmp.com/t5/Discussions/Malfunction-in-Bootstrap-Forest-with-Tuning-Design-Table/m-p/715269#U715269" discussion_style_icon_css="lia-mention-container-editor-message lia-img-icon-forum-thread lia-fa-icon lia-fa-forum lia-fa-thread lia-fa"&gt;&lt;/LI-MESSAGE&gt;&amp;nbsp;&lt;/LI&gt;
&lt;LI&gt;When launching the Boosted Tree (or any other modeling platform, specify your inputs, the response to model, and use the K-folds validation column in the validation panel (here on the Mushroom JMP dataset):&amp;nbsp;&lt;BR /&gt;&lt;span class="lia-inline-image-display-wrapper lia-image-align-inline" image-alt="Victor_G_0-1723533425820.png" style="width: 400px;"&gt;&lt;img src="https://community.jmp.com/t5/image/serverpage/image-id/67085i655C838EADAD3FCB/image-size/medium?v=v2&amp;amp;px=400" role="button" title="Victor_G_0-1723533425820.png" alt="Victor_G_0-1723533425820.png" /&gt;&lt;/span&gt;&lt;/LI&gt;
&lt;LI&gt;A new window pops up, you have to check "Use Tuning table" and then select your tuning table already open :&amp;nbsp;&lt;BR /&gt;&lt;span class="lia-inline-image-display-wrapper lia-image-align-inline" image-alt="Victor_G_1-1723533477035.png" style="width: 400px;"&gt;&lt;img src="https://community.jmp.com/t5/image/serverpage/image-id/67086i4DB602B40642BB9D/image-size/medium?v=v2&amp;amp;px=400" role="button" title="Victor_G_1-1723533477035.png" alt="Victor_G_1-1723533477035.png" /&gt;&lt;/span&gt;&lt;/LI&gt;
&lt;LI&gt;You'll then have the results provided by the best tuned models :&lt;BR /&gt;&lt;span class="lia-inline-image-display-wrapper lia-image-align-inline" image-alt="Victor_G_2-1723533617587.png" style="width: 400px;"&gt;&lt;img src="https://community.jmp.com/t5/image/serverpage/image-id/67087i6ABECFB8E5BCE7AF/image-size/medium?v=v2&amp;amp;px=400" role="button" title="Victor_G_2-1723533617587.png" alt="Victor_G_2-1723533617587.png" /&gt;&lt;/span&gt;&lt;/LI&gt;
&lt;/OL&gt;
&lt;P&gt;Note that this method is not suited for K-folds cross-validation, as the use of tuning tables imply a partition of your data only in 3 sets :&lt;/P&gt;
&lt;OL&gt;
&lt;LI&gt;Training set : Used for&amp;nbsp;&lt;SPAN&gt;the actual training of the model(s),&lt;/SPAN&gt;&lt;/LI&gt;
&lt;LI&gt;Validation set : Used for model optimization (hyperparameter fine-tuning, features/threshold selection, ... for example) and model selection,&lt;/LI&gt;
&lt;LI&gt;Test set : Used for generalization and predictive performance assessment of the selected model on new/unseen data.&lt;/LI&gt;
&lt;/OL&gt;
&lt;P&gt;So if you specify a 5-folds crossvalidation in step 1, only the first 3 folds will be used as training, validation and test sets, not really as a 5-folds crossvalidation technique.&amp;nbsp;To do a crossvalidation like you intend to do, you would need a nested crossvalidation : an inner crossvalidation to tune hyperpararameters, and an outer crossvalidation to assess robustness of the hyperparameters values found, only available in Model Screening platform and not accepting Tuning tables for hyperparameters tuning (as the goal of this platform is to screen the most promising algorithms among a large variety of model types, not finetune them) :&lt;BR /&gt;&lt;BR /&gt;&lt;/P&gt;
&lt;P&gt;&lt;span class="lia-inline-image-display-wrapper lia-image-align-inline" image-alt="Victor_G_3-1723534064034.png" style="width: 400px;"&gt;&lt;img src="https://community.jmp.com/t5/image/serverpage/image-id/67088iF9C4FB558FBE0708/image-size/medium?v=v2&amp;amp;px=400" role="button" title="Victor_G_3-1723534064034.png" alt="Victor_G_3-1723534064034.png" /&gt;&lt;/span&gt;&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;As far as I know, this is not (directly) possible in JMP.&lt;/P&gt;
&lt;P&gt;But you can still use K-folds crossvalidation on "default" Boosted Tree, or try using other validation techniques, following the method above (but creating a normal formula validation column with 3 sets) and using simulation on the tuned model to assess its robustness and benefit vs. a non-tuned one :&lt;BR /&gt;&lt;BR /&gt;&lt;/P&gt;
&lt;P&gt;&lt;span class="lia-inline-image-display-wrapper lia-image-align-inline" image-alt="Victor_G_4-1723535038590.png" style="width: 400px;"&gt;&lt;img src="https://community.jmp.com/t5/image/serverpage/image-id/67089i6D687CFE623ADF23/image-size/medium?v=v2&amp;amp;px=400" role="button" title="Victor_G_4-1723535038590.png" alt="Victor_G_4-1723535038590.png" /&gt;&lt;/span&gt;&lt;/P&gt;
&lt;P&gt;You can see that in most cases, the default hyperparameters values work quite well, and the hyperparameters tuning help more on the performances variability (performances values like RASE, R-square, ... have narrower ranges on the tuned algorithm compared to the "default" one) than on the max or average algorithm performances.&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;You can check a similar post and solution here on this topic to have a look at validation techniques and the use of simulation:&amp;nbsp;&lt;A href="https://community.jmp.com/t5/Discussions/Boosted-Tree-Tuning-TABLE-DESIGN/m-p/609591/highlight/true#M81062" target="_blank" rel="noopener"&gt;Solved: Re: Boosted Tree - Tuning TABLE DESIGN - JMP User Community&lt;/A&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;Nested cross-validation is typically not the first option I would recommend, as it requires a lot of computation, to fine-tune the algorithm independantly on each of the folds of the inner loop, and then calculate the performances on each validation fold of the outer loop. &lt;BR /&gt;K-folds crossvalidation is a useful technique in the absence of large quantity of data, but the nested cross-validation still requires a quite large amount to correctly do the data splitting: for example, if you split the inner loop in 4 folds and the outer loop in 5 folds, it requires to create 20 folds/groups in your dataset ! &lt;BR /&gt;Finally, crossvalidation is more a tuning technique than a validation technique to assess model robustness, as brillantly described by Cassie Kozyrkov in this video :&amp;nbsp;&lt;A href="https://youtu.be/zqD0lQy_w40?si=lja79_aik0KO-jbB" target="_blank"&gt;https://youtu.be/zqD0lQy_w40?si=lja79_aik0KO-jbB&lt;/A&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;Hope this answer will help you,&lt;/P&gt;</description>
      <pubDate>Tue, 13 Aug 2024 08:27:24 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/Is-there-a-way-to-do-k-fold-cross-validation-with-boosted-tree/m-p/781267#M96381</guid>
      <dc:creator>Victor_G</dc:creator>
      <dc:date>2024-08-13T08:27:24Z</dc:date>
    </item>
    <item>
      <title>Re: Is there a way to do k-fold cross validation with boosted tree?</title>
      <link>https://community.jmp.com/t5/Discussions/Is-there-a-way-to-do-k-fold-cross-validation-with-boosted-tree/m-p/782233#M96547</link>
      <description>&lt;P&gt;&lt;a href="https://community.jmp.com/t5/user/viewprofilepage/user-id/16391"&gt;@shampton82&lt;/a&gt;Thank you so much, this allowed me to create the k fold (val) column and run the boosted tree!&lt;/P&gt;</description>
      <pubDate>Fri, 16 Aug 2024 15:41:12 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/Is-there-a-way-to-do-k-fold-cross-validation-with-boosted-tree/m-p/782233#M96547</guid>
      <dc:creator>NishaKumar2023</dc:creator>
      <dc:date>2024-08-16T15:41:12Z</dc:date>
    </item>
    <item>
      <title>Re: Is there a way to do k-fold cross validation with boosted tree?</title>
      <link>https://community.jmp.com/t5/Discussions/Is-there-a-way-to-do-k-fold-cross-validation-with-boosted-tree/m-p/782234#M96548</link>
      <description>&lt;P&gt;Thank you so much &lt;a href="https://community.jmp.com/t5/user/viewprofilepage/user-id/11568"&gt;@Victor_G&lt;/a&gt; ! I found your explanation to be very help esp for the next step in hypertuning, I was able to create the k fold val column via xgboost and run the boosted tree but I will directions and resources to be helpful for the next few steps. Thank you!&lt;/P&gt;</description>
      <pubDate>Fri, 16 Aug 2024 15:43:19 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/Is-there-a-way-to-do-k-fold-cross-validation-with-boosted-tree/m-p/782234#M96548</guid>
      <dc:creator>NishaKumar2023</dc:creator>
      <dc:date>2024-08-16T15:43:19Z</dc:date>
    </item>
  </channel>
</rss>

