<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic model screening in Discussions</title>
    <link>https://community.jmp.com/t5/Discussions/model-screening/m-p/752105#M93347</link>
    <description>&lt;P&gt;Hello.&lt;BR /&gt;The question I have is that I have a dataset, and I want to separate 30% of this dataset for testing. I want to try different models using model screening, but for the validation part of my models, I want all my models to use exactly the same test dataset for validation.&lt;BR /&gt;What should I do for this purpose?&lt;/P&gt;</description>
    <pubDate>Sun, 05 May 2024 10:23:40 GMT</pubDate>
    <dc:creator>maryam_nourmand</dc:creator>
    <dc:date>2024-05-05T10:23:40Z</dc:date>
    <item>
      <title>model screening</title>
      <link>https://community.jmp.com/t5/Discussions/model-screening/m-p/752105#M93347</link>
      <description>&lt;P&gt;Hello.&lt;BR /&gt;The question I have is that I have a dataset, and I want to separate 30% of this dataset for testing. I want to try different models using model screening, but for the validation part of my models, I want all my models to use exactly the same test dataset for validation.&lt;BR /&gt;What should I do for this purpose?&lt;/P&gt;</description>
      <pubDate>Sun, 05 May 2024 10:23:40 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/model-screening/m-p/752105#M93347</guid>
      <dc:creator>maryam_nourmand</dc:creator>
      <dc:date>2024-05-05T10:23:40Z</dc:date>
    </item>
    <item>
      <title>Re: model screening</title>
      <link>https://community.jmp.com/t5/Discussions/model-screening/m-p/752120#M93348</link>
      <description>&lt;P&gt;Hi&amp;nbsp;&lt;a href="https://community.jmp.com/t5/user/viewprofilepage/user-id/56938"&gt;@maryam_nourmand&lt;/a&gt;,&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;The easiest way to have reproducible results on fixed training/validation/test datasets is to&amp;nbsp;&lt;A href="https://www.jmp.com/support/help/en/17.2/#page/jmp/make-validation-column.shtml#" target="_blank" rel="noopener"&gt;Make Validation Column.&lt;/A&gt;&lt;/P&gt;
&lt;OL&gt;
&lt;LI&gt;First, go into the corresponding menu for creating validation column :&lt;BR /&gt;&lt;span class="lia-inline-image-display-wrapper lia-image-align-inline" image-alt="Capture d'écran 2024-05-05 124224.png" style="width: 400px;"&gt;&lt;img src="https://community.jmp.com/t5/image/serverpage/image-id/63923iB0E4A583913814C3/image-size/medium?v=v2&amp;amp;px=400" role="button" title="Capture d'écran 2024-05-05 124224.png" alt="Capture d'écran 2024-05-05 124224.png" /&gt;&lt;/span&gt;&lt;/LI&gt;
&lt;LI&gt;Then you can specify groups or stratification (if needed) :&lt;BR /&gt;&lt;span class="lia-inline-image-display-wrapper lia-image-align-inline" image-alt="Victor_G_0-1714905910194.png" style="width: 400px;"&gt;&lt;img src="https://community.jmp.com/t5/image/serverpage/image-id/63924iAECDBD636806834E/image-size/medium?v=v2&amp;amp;px=400" role="button" title="Victor_G_0-1714905910194.png" alt="Victor_G_0-1714905910194.png" /&gt;&lt;/span&gt;
&lt;P&gt;You have several options for the method: here I used the default one corresponding to your needs, "Make Validation Column" and I use some variables as stratification columns to make sure all my training, validation and test sets will have a similar distribution/repartition of values for these columns.&lt;/P&gt;
&lt;/LI&gt;
&lt;LI&gt;
&lt;P&gt;Then you can specify the ratio of data for each of your sets, and specify a random seed to have perfect reproducibility in your data splitting. Here I choosed a fixed validation type, but you can also create a formula validation column, enabling to simulate various training, validation and test sets with the same settings (grouping, stratification, etc...) as specified in step 2. Here my data ratios are 70% for training, 20% for validation, and 10% for test (usual values, for more info and reasearch on this topic, you can read the paper "Optimal Ratio for Data Splitting" by V. Roshan Joseph: &lt;A href="https://arxiv.org/pdf/2202.03326" target="_blank" rel="noopener"&gt;https://arxiv.org/pdf/2202.03326&lt;/A&gt;). It's important to find a good compromise for the ratios depending on the assessment accuracy/fairness and model selection needed, and the number of data/rows available. When you're done, click on "Go" to create this new validation column in your table :&lt;/P&gt;
&lt;span class="lia-inline-image-display-wrapper lia-image-align-inline" image-alt="Victor_G_1-1714906109803.png" style="width: 400px;"&gt;&lt;img src="https://community.jmp.com/t5/image/serverpage/image-id/63925iF7EE6281F27421F7/image-size/medium?v=v2&amp;amp;px=400" role="button" title="Victor_G_1-1714906109803.png" alt="Victor_G_1-1714906109803.png" /&gt;&lt;/span&gt;&lt;/LI&gt;
&lt;LI&gt;
&lt;P&gt;Finally, launch the Model screening platform and use your newly created "Validation column" in the "Validation" panel. You can also fix a random seed at this stage to have reproducible results for the different models and sets used :&lt;/P&gt;
&lt;span class="lia-inline-image-display-wrapper lia-image-align-inline" image-alt="Victor_G_2-1714906346339.png" style="width: 400px;"&gt;&lt;img src="https://community.jmp.com/t5/image/serverpage/image-id/63926i947BC7F72B690667/image-size/medium?v=v2&amp;amp;px=400" role="button" title="Victor_G_2-1714906346339.png" alt="Victor_G_2-1714906346339.png" /&gt;&lt;/span&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;/LI&gt;
&lt;/OL&gt;
&lt;P&gt;Just some clarifications about the terms validation and test, as they may very often be used alternatively with some confusion :&lt;/P&gt;
&lt;UL&gt;
&lt;LI&gt;Training set : Used for&amp;nbsp;&lt;SPAN&gt;the actual training of the model(s),&lt;/SPAN&gt;&lt;/LI&gt;
&lt;LI&gt;Validation set : Used for model optimization (hyperparameter fine-tuning, features/threshold selection, ... for example) and model selection,&lt;/LI&gt;
&lt;LI&gt;Test set : Used for generalization and predictive performance assessment of the selected model on new/unseen data.&lt;/LI&gt;
&lt;/UL&gt;
&lt;P&gt;So you can evaluate several models with the validation set, but the test set should be used only once for the model selected in the validation phase.&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;Hope this answer will help you,&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt; &lt;/P&gt;</description>
      <pubDate>Sun, 05 May 2024 11:09:22 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/model-screening/m-p/752120#M93348</guid>
      <dc:creator>Victor_G</dc:creator>
      <dc:date>2024-05-05T11:09:22Z</dc:date>
    </item>
    <item>
      <title>Re: model screening</title>
      <link>https://community.jmp.com/t5/Discussions/model-screening/m-p/752121#M93349</link>
      <description>&lt;P&gt;thanks&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Sun, 05 May 2024 11:25:23 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/model-screening/m-p/752121#M93349</guid>
      <dc:creator>maryam_nourmand</dc:creator>
      <dc:date>2024-05-05T11:25:23Z</dc:date>
    </item>
    <item>
      <title>Re: model screening</title>
      <link>https://community.jmp.com/t5/Discussions/model-screening/m-p/752122#M93350</link>
      <description>&lt;P&gt;If you have JMP Pro it has a platform to create a Validation Column, and within the various modeling platforms one can specify the column that indicates which rows are Training rows and which are Validation rows.&lt;/P&gt;
&lt;P&gt;If you have standard JMP, you can easily create a Validation column by using the&amp;nbsp; Random Uniform() function in the formula for the Validation column being created. (Make sure you specify the column as being a Character Column)&lt;/P&gt;
&lt;PRE&gt;&lt;CODE class=" language-jsl"&gt;If(randome uniform()&amp;gt;=.7, "Validation", "Training")&lt;/CODE&gt;&lt;/PRE&gt;
&lt;P&gt;After the column is created, you will want to go to the Column Info screen, and remove the formula which will make the column's values static, insuring that the formula is not run a second time, changing the values.&lt;/P&gt;
&lt;P&gt;You can then Exclude and Hide the Validation rows while you do your model building, and then when you want to validate the mode, un Hide and Exclude the Validation rows and Hide and Exclude the Training rows and then rerun the model.&amp;nbsp;&amp;nbsp;&lt;/P&gt;
&lt;P class="N1bullet"&gt;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Sun, 05 May 2024 11:29:45 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/model-screening/m-p/752122#M93350</guid>
      <dc:creator>txnelson</dc:creator>
      <dc:date>2024-05-05T11:29:45Z</dc:date>
    </item>
  </channel>
</rss>

