<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Re: SVEM ADD IN FOR SUPPORT VECTOR MACHINE in Discussions</title>
    <link>https://community.jmp.com/t5/Discussions/SVEM-ADD-IN-FOR-SUPPORT-VECTOR-MACHINE/m-p/767310#M94750</link>
    <description>&lt;P&gt;Hi&amp;nbsp;&lt;a href="https://community.jmp.com/t5/user/viewprofilepage/user-id/283"&gt;@frankderuyck&lt;/a&gt;,&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;This is two different considerations :&lt;/P&gt;
&lt;UL&gt;
&lt;LI&gt;SVEM is a validation strategy that could be applied to any model, from linear regression to neural networks.&lt;/LI&gt;
&lt;LI&gt;Random Forest is a Machine Learning model.&lt;/LI&gt;
&lt;/UL&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;Both share in common a way to prevent overfitting when modeling :&lt;/P&gt;
&lt;UL&gt;
&lt;LI&gt;Random Forest use bootstrap samples and "&lt;A href="https://www.jmp.com/support/help/en/17.2/#page/jmp/per-tree-summaries.shtml" target="_self"&gt;Out-of-Bag&lt;/A&gt;" sample as "internal" model validation. The "Out-of-Bag" sample is a part of the data not used in training: Training samples are bootstrap samples sampled with replacement in the original dataset. Theoritically, you can calculate that by sampling with replacement your original data, around 1/3 of your data won't be sampled :&amp;nbsp;&lt;A href="https://stats.stackexchange.com/questions/173520/random-forests-out-of-bag-sample-size" target="_blank" rel="noopener"&gt;machine learning - Random Forests out-of-bag sample size - Cross Validated (stackexchange.com)&lt;/A&gt;&amp;nbsp;&lt;BR /&gt;This sample constitutes the "Out-of-bag" sample not used in the training of the Random Forest and enable to assess/validate the relevance and precision of the model.&lt;/LI&gt;
&lt;LI&gt;SVEM use anticorrelated weights for training and validation, meaning that an experiment with a high weight for training will have a low weight for validation. By fitting a model with this validation setup, and then changing the weights and refitting the model with new weights for training and validation a high number of times, you have slightly different models that you can ensemble/combine, in order to reduce variance (and prevent overfitting).&amp;nbsp;&lt;/LI&gt;
&lt;/UL&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;Now, my &lt;STRONG&gt;personal&lt;/STRONG&gt; opinion and choice :&lt;/P&gt;
&lt;UL&gt;
&lt;LI&gt;What I do see frequently is people using SVEM for their DoE with highly complex models without even further considerations or previous analysis/modeling. You can often see for example SVEM applied with Neural Networks for quite relatively simple designs and results.&amp;nbsp;It sounds like "&lt;EM&gt;using a bazooka to kill a fly&lt;/EM&gt;" for me : I tend to choose the most simple options/models before increasing the complexity (if necessary) : &lt;EM&gt;Ockham's razor&lt;/EM&gt;. You also have to consider how much levels you have for your factors and other considerations (like dimensionality), as Machine Learning models are interpolating models, very efficient at finding a pattern between points. So if you only have 2-3 levels for your factors, that might be not enough to really gain a benefit using Machine Learning.&lt;/LI&gt;
&lt;LI&gt;If you have enough levels and the traditional linear modeling doesn't seem appropriate because of non-linear relationships, then you might be interested to try Machine Learning models.&lt;/LI&gt;
&lt;LI&gt;Since you have a low-size but high-quality training dataset with your DoE, you have to choose a ML model that is simple (no or few and easy hyperparameters fine-tuning), robust (less sensitive to hyperparameters values setting) and less prone to overfitting, since you can't split your DoE data into a training and validation sets. This is where Random Forests (with its internal "Out-of-bag" sample validation) and SVM (you can evaluate overfitting possibility with hyperparameters choice and number of support vectors) are interesting.&amp;nbsp;&lt;/LI&gt;
&lt;/UL&gt;
&lt;P&gt;Also using a robust ML algorithm instead of a complex one with a SVEM strategy has other benefits, such as lower computational time and easier interpretability.&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;Hope this answer will help you and make sense for you,&lt;/P&gt;</description>
    <pubDate>Thu, 20 Jun 2024 12:39:30 GMT</pubDate>
    <dc:creator>Victor_G</dc:creator>
    <dc:date>2024-06-20T12:39:30Z</dc:date>
    <item>
      <title>SVEM ADD IN FOR SUPPORT VECTOR MACHINE</title>
      <link>https://community.jmp.com/t5/Discussions/SVEM-ADD-IN-FOR-SUPPORT-VECTOR-MACHINE/m-p/762835#M94296</link>
      <description>&lt;P&gt;I have a DOE results I would like to model with Support Vector Machine (there are strong non linear effects that can't be estimated with a polynomial model) How do I install the SVEM add in?&lt;/P&gt;</description>
      <pubDate>Fri, 07 Jun 2024 07:29:34 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/SVEM-ADD-IN-FOR-SUPPORT-VECTOR-MACHINE/m-p/762835#M94296</guid>
      <dc:creator>frankderuyck</dc:creator>
      <dc:date>2024-06-07T07:29:34Z</dc:date>
    </item>
    <item>
      <title>Re: SVEM ADD IN FOR SUPPORT VECTOR MACHINE</title>
      <link>https://community.jmp.com/t5/Discussions/SVEM-ADD-IN-FOR-SUPPORT-VECTOR-MACHINE/m-p/762840#M94298</link>
      <description>&lt;P&gt;Hi&amp;nbsp;&lt;a href="https://community.jmp.com/t5/user/viewprofilepage/user-id/283"&gt;@frankderuyck&lt;/a&gt;,&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;Some questions before answering yours :&lt;/P&gt;
&lt;UL&gt;
&lt;LI&gt;&lt;STRONG&gt;&lt;EM&gt;Why would you like to use SVEM for SVM ?&lt;/EM&gt;&lt;/STRONG&gt; Support Vector Machines is a robust ML algorithm, that can be used for small-sized datasets like DoE. More infos about algorithm's robustness on&amp;nbsp;&lt;A href="https://www.linkedin.com/posts/victorguiller_doe-machinelearning-statsjoke-activity-7117395916130537472-pJoB/" target="_blank" rel="noopener"&gt;LinkedIn.&lt;/A&gt;&lt;BR /&gt;There are also many ways to assess overfitting for SVM without using validation set, by checking the number of support vectors, and the values of penalty (cost) and curvature (gamma) if you have fine-tuned your SVM hyperparameters with a &lt;A href="https://www.jmp.com/support/help/en/17.2/#page/jmp/model-launch-control-panel-2.shtml#" target="_self"&gt;Tuning design&lt;/A&gt;... I am thinking about writing a blog post about assessing and preventing overfitting for robust algorithms (SVM, Random Forest) in the absence of validation (in DoE datasets for example).&lt;/LI&gt;
&lt;LI&gt;&lt;STRONG&gt;&lt;EM&gt;Did you also consider other algorithms ?&lt;/EM&gt;&lt;/STRONG&gt;&amp;nbsp;SVM is a robust algorithm for analysis of small-size datasets, but Random Forest can also provide good results and may be a good benchmark to evaluate and compare performances. More infos on Random Forests on&amp;nbsp;&lt;A href="https://www.linkedin.com/posts/victorguiller_doe-machinelearning-randomforests-activity-7127557799810625536-FH7K/?trk=public_profile_like_view" target="_blank" rel="noopener"&gt;LinkedIn.&lt;/A&gt;&lt;/LI&gt;
&lt;LI&gt;&lt;STRONG&gt;&lt;EM&gt;How many levels do you have for your factors ?&lt;/EM&gt;&lt;/STRONG&gt; ML algorithms can only be useful if you have enough levels for each of your factors to enable a smooth interpolation. More than 3 levels may be a good starting point.&lt;/LI&gt;
&lt;/UL&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;Concerning the use of SVEM, you can access it through different ways :&lt;/P&gt;
&lt;UL&gt;
&lt;LI&gt;In the platform "Make Validation Column", you have the possibility to create an Autovalidation table :&amp;nbsp;&lt;A href="https://www.jmp.com/support/help/en/17.2/#page/jmp/launch-the-make-validation-column-platform.shtml" target="_blank" rel="noopener"&gt;Launch the Make Validation Column Platform (jmp.com)&lt;/A&gt;&amp;nbsp;This will duplicate the rows and create anti-correlated weights (between training and validation) column that you can use as "Frequency" in the platform launch. You can then simulate prediction results by varying the weights between training and validation (right-click a panel with validation results, click on Simulate, and specify weights column as the column to switch in and out).&lt;/LI&gt;
&lt;LI&gt;You can also download the addin SVEM, that will do the same things as the autovalidation table :&amp;nbsp;&lt;A href="https://community.jmp.com/t5/Discovery-Summit-Europe-2021/Re-Thinking-the-Design-and-Analysis-of-Experiments-2021-EU-30MP/ta-p/349240#U349240" target="_blank" rel="noopener"&gt;Re-Thinking the Design and Analysis of Experiments? (2021-EU-30MP-776) - JMP User Community&lt;/A&gt;&lt;BR /&gt;You can then reproduce the same steps as before and simulate results to assess model's robustness.&lt;/LI&gt;
&lt;/UL&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;I would however like to emphasize that cross-validation (K-folds or SVEM) is not properly a validation method, but more a "debugging" method to assess robustness of algorithm's performances on various training and validation sets. The main difficulty after using a cross-validation technique is how to make sense of the results and how to proceed next:&amp;nbsp;&lt;BR /&gt;Average the individual models trained on various folds ? Create an ensemble with these models ? Keep the best model trained on a fold (I wouldn't recommend this option) ? Train the model on whole dataset once the robustness has been checked with the cross-validation ? ... There is no definitive answer to this question and this is related to the dataset, modelisation complexity, performances improvement, user's practices, ...&lt;BR /&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;More infos here with some episodes of Cassie Kozyrkov on the topic of cross-validation:&amp;nbsp;&lt;A href="https://youtu.be/zqD0lQy_w40?si=41jfBs83Q0TIlusr" target="_blank" rel="noopener"&gt;https://youtu.be/zqD0lQy_w40?si=41jfBs83Q0TIlusr&lt;/A&gt;&lt;/P&gt;
&lt;P&gt;&lt;A href="https://youtu.be/by59hsYO0Lg?si=XeZ33j2e4jKQPaRA" target="_blank" rel="noopener"&gt;https://youtu.be/by59hsYO0Lg?si=XeZ33j2e4jKQPaRA&lt;/A&gt;&amp;nbsp;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;Hope this answer will help you,&lt;/P&gt;</description>
      <pubDate>Fri, 07 Jun 2024 08:31:09 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/SVEM-ADD-IN-FOR-SUPPORT-VECTOR-MACHINE/m-p/762840#M94298</guid>
      <dc:creator>Victor_G</dc:creator>
      <dc:date>2024-06-07T08:31:09Z</dc:date>
    </item>
    <item>
      <title>Re: SVEM ADD IN FOR SUPPORT VECTOR MACHINE</title>
      <link>https://community.jmp.com/t5/Discussions/SVEM-ADD-IN-FOR-SUPPORT-VECTOR-MACHINE/m-p/762853#M94306</link>
      <description>&lt;P&gt;Hi Victor, this tuning design works great,&amp;nbsp; I want to model a guality response Y as a function of three Y FPCA's from NIR spectra (FDA). I tried neural network but even with 3 TanH nodes profiler shows to my opinion too high and irregular ncurvature. SVE after tuning does a nice job,&amp;nbsp;I support vector machine learning!&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&lt;span class="lia-inline-image-display-wrapper lia-image-align-inline" image-alt="frankderuyck_0-1717763077556.png" style="width: 308px;"&gt;&lt;img src="https://community.jmp.com/t5/image/serverpage/image-id/64985i468C50E6311E72D7/image-dimensions/308x229?v=v2" width="308" height="229" role="button" title="frankderuyck_0-1717763077556.png" alt="frankderuyck_0-1717763077556.png" /&gt;&lt;/span&gt;&lt;/P&gt;&lt;P&gt;&lt;span class="lia-inline-image-display-wrapper lia-image-align-inline" image-alt="frankderuyck_1-1717763142689.png" style="width: 400px;"&gt;&lt;img src="https://community.jmp.com/t5/image/serverpage/image-id/64986i35A7155ECCB63BC7/image-size/medium?v=v2&amp;amp;px=400" role="button" title="frankderuyck_1-1717763142689.png" alt="frankderuyck_1-1717763142689.png" /&gt;&lt;/span&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Fri, 07 Jun 2024 12:26:18 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/SVEM-ADD-IN-FOR-SUPPORT-VECTOR-MACHINE/m-p/762853#M94306</guid>
      <dc:creator>frankderuyck</dc:creator>
      <dc:date>2024-06-07T12:26:18Z</dc:date>
    </item>
    <item>
      <title>Re: SVEM ADD IN FOR SUPPORT VECTOR MACHINE</title>
      <link>https://community.jmp.com/t5/Discussions/SVEM-ADD-IN-FOR-SUPPORT-VECTOR-MACHINE/m-p/762856#M94308</link>
      <description>&lt;P&gt;Hi&amp;nbsp;&lt;a href="https://community.jmp.com/t5/user/viewprofilepage/user-id/283"&gt;@frankderuyck&lt;/a&gt;,&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;Great to know ! :)&lt;/img&gt;&lt;BR /&gt;I use Neural Networks as the last modeling option possible, there are many classical models and ML algorithms more simple that perform equally well (and sometimes better). These other options are also faster to compute, don't require extended knowledge, tests and validation to finetune them (unlike NNs where you have to specify the architecture, with the number of layers, type of functions, boosting or not, ... or using the &lt;A href="https://community.jmp.com/t5/JMP-Add-Ins/Neural-Network-Tuning/ta-p/662666" target="_blank"&gt;Neural Network Tuning AddIn&lt;/A&gt;) and are easier to interpret (in terms of response profiles and relative factors importance). &lt;BR /&gt;On tabular data, if the only model able to perform well is Neural Networks, there are good chances that you might have other problems, in terms of data quality, representativeness, missing values/outliers, repeatability/reproducibility, etc...&lt;/P&gt;
&lt;P&gt;Tree-based models provide a good performances benchmark and are often a safe and reliable option, see the article "&lt;STRONG&gt;Why do tree-based models still outperform deep learning on tabular data&lt;/STRONG&gt;?" (2022) &lt;A href="https://arxiv.org/abs/2207.08815" target="_blank"&gt;https://arxiv.org/abs/2207.08815&lt;/A&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;In your screenshot, it seems the optimal SVM model found with the Tuning design is quite interesting :&lt;/P&gt;
&lt;UL&gt;
&lt;LI&gt;It uses 10 support vectors, so less than half of the points are used to build the "boundaries"/vectors used in SVM model, which may be a good sign of low complexity and good generalization properties,&lt;/LI&gt;
&lt;LI&gt;The curvature value (gamma) is quite low, so the response profile and boundaries may not be over-complex and highly curved (so again, low complexity and good generalization properties),&lt;/LI&gt;
&lt;LI&gt;The penalty (cost value) seems quite high, but with a low-size dataset this is often the case in practice, as any misclassified/mispredicted point will greatly increase the misclassification rate/RMSE : 1 misclassified point out of 25 = 4% misclassification rate, compared to higher dataset size where one misclassified point would represent less than 1% misclassification rate. In some situations, it can be interesting to choose a model with a slightly larger error (RMSE/Misclassification rate), but with a lot less support vectors : you degrade your predictive performances on the training set in favor of a less complex model that could be able to generalize better on new data.&lt;/LI&gt;
&lt;/UL&gt;
&lt;P&gt;You can use the table generated by the Tuning design, with the performance metrics, to better assess the relative importance and contribution of the two hyperparameters on the performance metrics (Misclassification rate for classification, or RASE and R² for regression).&amp;nbsp;&lt;/P&gt;
&lt;P&gt;As always, it is recommended to validate your model with new validation points, to make sure your model doesn't overfit.&lt;/P&gt;</description>
      <pubDate>Fri, 07 Jun 2024 13:07:27 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/SVEM-ADD-IN-FOR-SUPPORT-VECTOR-MACHINE/m-p/762856#M94308</guid>
      <dc:creator>Victor_G</dc:creator>
      <dc:date>2024-06-07T13:07:27Z</dc:date>
    </item>
    <item>
      <title>Re: SVEM ADD IN FOR SUPPORT VECTOR MACHINE</title>
      <link>https://community.jmp.com/t5/Discussions/SVEM-ADD-IN-FOR-SUPPORT-VECTOR-MACHINE/m-p/762867#M94312</link>
      <description>&lt;P&gt;Guess tree based models like random forest &amp;amp; XGboost require a large data set unless you can use SVEM for small sets, correct?&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Fri, 07 Jun 2024 14:21:34 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/SVEM-ADD-IN-FOR-SUPPORT-VECTOR-MACHINE/m-p/762867#M94312</guid>
      <dc:creator>frankderuyck</dc:creator>
      <dc:date>2024-06-07T14:21:34Z</dc:date>
    </item>
    <item>
      <title>Re: SVEM ADD IN FOR SUPPORT VECTOR MACHINE</title>
      <link>https://community.jmp.com/t5/Discussions/SVEM-ADD-IN-FOR-SUPPORT-VECTOR-MACHINE/m-p/762873#M94314</link>
      <description>&lt;P&gt;I don't fully agree, there are more nuances between tree-based models:&lt;/P&gt;
&lt;UL&gt;
&lt;LI&gt;&lt;EM&gt;Boosted Tree methods&lt;/EM&gt;, like XGBoost and others, use trees trained sequentially to refine and correct predictions progressively. They are more influenced by the training set characteristics, size, data quality, repartition of data points, representativeness, ... as they iteratively try to have the best predictions (each tree "correcting" the largest prediction errors of the previous tree) on the dataset they are trained on. &lt;BR /&gt;It's typically an algorithm described as low bias - high variance : predictions can be very precise, but very sensitive/influenced by the training dataset used.&lt;BR /&gt;&lt;BR /&gt;&lt;/LI&gt;
&lt;LI&gt;&lt;EM&gt;Random Forests&lt;/EM&gt; use trees trained in parallel on slightly different datasets (bootstrap sets). The good predictive performance of this algorithm comes from the averaging of the individual trees predictions. You can think of it as "&lt;EM&gt;&lt;STRONG&gt;T&lt;/STRONG&gt;&lt;STRONG&gt;he Wisdom of the Crowd&lt;/STRONG&gt;&lt;/EM&gt;": if you ask the weight of a whale to someone, it might be very inaccurate and not precise, but if collect the answers from hundreds or thousands of people and average the results, the mean result may be very close to the actual value.&amp;nbsp; &lt;BR /&gt;It's typically an algorithm described as medium/low bias - low variance : predictions may be precise (depending on the averaging of the individual results and the complexity/number of individual trees), and they are quite robust to the training dataset thanks to the bootstrapping process and the individual training of the trees (with different features).&lt;/LI&gt;
&lt;/UL&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;Here is a short visual between the different tree-based models (I don't like how the individual trees in the Random Forest have the same structure, but it's a simplistic schema):&lt;/P&gt;
&lt;P&gt;&lt;span class="lia-inline-image-display-wrapper lia-image-align-inline" image-alt="DecisionTree-BoostedTree-RandomForest.png" style="width: 400px;"&gt;&lt;img src="https://community.jmp.com/t5/image/serverpage/image-id/64989iA0713E8567DDBA29/image-size/medium?v=v2&amp;amp;px=400" role="button" title="DecisionTree-BoostedTree-RandomForest.png" alt="DecisionTree-BoostedTree-RandomForest.png" /&gt;&lt;/span&gt;&lt;/P&gt;
&lt;P&gt;To answer your question, I would try Random Forests over Boosted Tree methods on low size dataset like DoE, to have more confidence in the results, use an algorithm less prone to overfitting, and have better generalization properties.&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Fri, 07 Jun 2024 15:25:11 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/SVEM-ADD-IN-FOR-SUPPORT-VECTOR-MACHINE/m-p/762873#M94314</guid>
      <dc:creator>Victor_G</dc:creator>
      <dc:date>2024-06-07T15:25:11Z</dc:date>
    </item>
    <item>
      <title>Re: SVEM ADD IN FOR SUPPORT VECTOR MACHINE</title>
      <link>https://community.jmp.com/t5/Discussions/SVEM-ADD-IN-FOR-SUPPORT-VECTOR-MACHINE/m-p/762879#M94317</link>
      <description>&lt;P&gt;By Random forest I get poor R² = 0,63 even with 2000 trees&lt;/P&gt;</description>
      <pubDate>Fri, 07 Jun 2024 15:40:47 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/SVEM-ADD-IN-FOR-SUPPORT-VECTOR-MACHINE/m-p/762879#M94317</guid>
      <dc:creator>frankderuyck</dc:creator>
      <dc:date>2024-06-07T15:40:47Z</dc:date>
    </item>
    <item>
      <title>Re: SVEM ADD IN FOR SUPPORT VECTOR MACHINE</title>
      <link>https://community.jmp.com/t5/Discussions/SVEM-ADD-IN-FOR-SUPPORT-VECTOR-MACHINE/m-p/762880#M94318</link>
      <description>&lt;P&gt;With boosted tree R² = 0,85&lt;/P&gt;</description>
      <pubDate>Fri, 07 Jun 2024 15:43:05 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/SVEM-ADD-IN-FOR-SUPPORT-VECTOR-MACHINE/m-p/762880#M94318</guid>
      <dc:creator>frankderuyck</dc:creator>
      <dc:date>2024-06-07T15:43:05Z</dc:date>
    </item>
    <item>
      <title>Re: SVEM ADD IN FOR SUPPORT VECTOR MACHINE</title>
      <link>https://community.jmp.com/t5/Discussions/SVEM-ADD-IN-FOR-SUPPORT-VECTOR-MACHINE/m-p/762881#M94319</link>
      <description>&lt;P&gt;Yes, boosted tree may be more precise than Random Forests "by design", due to their iterative prediction error correction with sequential trees training.&amp;nbsp;But the complex splits done during the sequential trees learning to improve the performance may only be valid on the training set, and may not reflect what could happen on new data points.&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;Random Forest is a robust algorithm regarding the hyperparameters, so increasing the number of trees above a certain value may not change the performance evaluation, see my post on&amp;nbsp;&lt;A href="https://www.linkedin.com/posts/victorguiller_design-of-experiments-machine-learning-activity-7122469274148868096-Qhif/" target="_blank"&gt;LinkedIn&lt;/A&gt;&amp;nbsp;and/or the article "&lt;EM&gt;Tunability: Importance of Hyperparameters of Machine Learning Algorithms&lt;/EM&gt;" :&amp;nbsp;&lt;A href="https://arxiv.org/pdf/1802.09596" target="_blank"&gt;1802.09596 (arxiv.org)&lt;/A&gt;&lt;/P&gt;</description>
      <pubDate>Fri, 07 Jun 2024 15:52:56 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/SVEM-ADD-IN-FOR-SUPPORT-VECTOR-MACHINE/m-p/762881#M94319</guid>
      <dc:creator>Victor_G</dc:creator>
      <dc:date>2024-06-07T15:52:56Z</dc:date>
    </item>
    <item>
      <title>Re: SVEM ADD IN FOR SUPPORT VECTOR MACHINE</title>
      <link>https://community.jmp.com/t5/Discussions/SVEM-ADD-IN-FOR-SUPPORT-VECTOR-MACHINE/m-p/767297#M94748</link>
      <description>&lt;P&gt;Hi Victor, with some delay... why would you choose Random Forest over SVEM for modeling DOE data?&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Thu, 20 Jun 2024 11:16:59 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/SVEM-ADD-IN-FOR-SUPPORT-VECTOR-MACHINE/m-p/767297#M94748</guid>
      <dc:creator>frankderuyck</dc:creator>
      <dc:date>2024-06-20T11:16:59Z</dc:date>
    </item>
    <item>
      <title>Re: SVEM ADD IN FOR SUPPORT VECTOR MACHINE</title>
      <link>https://community.jmp.com/t5/Discussions/SVEM-ADD-IN-FOR-SUPPORT-VECTOR-MACHINE/m-p/767310#M94750</link>
      <description>&lt;P&gt;Hi&amp;nbsp;&lt;a href="https://community.jmp.com/t5/user/viewprofilepage/user-id/283"&gt;@frankderuyck&lt;/a&gt;,&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;This is two different considerations :&lt;/P&gt;
&lt;UL&gt;
&lt;LI&gt;SVEM is a validation strategy that could be applied to any model, from linear regression to neural networks.&lt;/LI&gt;
&lt;LI&gt;Random Forest is a Machine Learning model.&lt;/LI&gt;
&lt;/UL&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;Both share in common a way to prevent overfitting when modeling :&lt;/P&gt;
&lt;UL&gt;
&lt;LI&gt;Random Forest use bootstrap samples and "&lt;A href="https://www.jmp.com/support/help/en/17.2/#page/jmp/per-tree-summaries.shtml" target="_self"&gt;Out-of-Bag&lt;/A&gt;" sample as "internal" model validation. The "Out-of-Bag" sample is a part of the data not used in training: Training samples are bootstrap samples sampled with replacement in the original dataset. Theoritically, you can calculate that by sampling with replacement your original data, around 1/3 of your data won't be sampled :&amp;nbsp;&lt;A href="https://stats.stackexchange.com/questions/173520/random-forests-out-of-bag-sample-size" target="_blank" rel="noopener"&gt;machine learning - Random Forests out-of-bag sample size - Cross Validated (stackexchange.com)&lt;/A&gt;&amp;nbsp;&lt;BR /&gt;This sample constitutes the "Out-of-bag" sample not used in the training of the Random Forest and enable to assess/validate the relevance and precision of the model.&lt;/LI&gt;
&lt;LI&gt;SVEM use anticorrelated weights for training and validation, meaning that an experiment with a high weight for training will have a low weight for validation. By fitting a model with this validation setup, and then changing the weights and refitting the model with new weights for training and validation a high number of times, you have slightly different models that you can ensemble/combine, in order to reduce variance (and prevent overfitting).&amp;nbsp;&lt;/LI&gt;
&lt;/UL&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;Now, my &lt;STRONG&gt;personal&lt;/STRONG&gt; opinion and choice :&lt;/P&gt;
&lt;UL&gt;
&lt;LI&gt;What I do see frequently is people using SVEM for their DoE with highly complex models without even further considerations or previous analysis/modeling. You can often see for example SVEM applied with Neural Networks for quite relatively simple designs and results.&amp;nbsp;It sounds like "&lt;EM&gt;using a bazooka to kill a fly&lt;/EM&gt;" for me : I tend to choose the most simple options/models before increasing the complexity (if necessary) : &lt;EM&gt;Ockham's razor&lt;/EM&gt;. You also have to consider how much levels you have for your factors and other considerations (like dimensionality), as Machine Learning models are interpolating models, very efficient at finding a pattern between points. So if you only have 2-3 levels for your factors, that might be not enough to really gain a benefit using Machine Learning.&lt;/LI&gt;
&lt;LI&gt;If you have enough levels and the traditional linear modeling doesn't seem appropriate because of non-linear relationships, then you might be interested to try Machine Learning models.&lt;/LI&gt;
&lt;LI&gt;Since you have a low-size but high-quality training dataset with your DoE, you have to choose a ML model that is simple (no or few and easy hyperparameters fine-tuning), robust (less sensitive to hyperparameters values setting) and less prone to overfitting, since you can't split your DoE data into a training and validation sets. This is where Random Forests (with its internal "Out-of-bag" sample validation) and SVM (you can evaluate overfitting possibility with hyperparameters choice and number of support vectors) are interesting.&amp;nbsp;&lt;/LI&gt;
&lt;/UL&gt;
&lt;P&gt;Also using a robust ML algorithm instead of a complex one with a SVEM strategy has other benefits, such as lower computational time and easier interpretability.&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;Hope this answer will help you and make sense for you,&lt;/P&gt;</description>
      <pubDate>Thu, 20 Jun 2024 12:39:30 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/SVEM-ADD-IN-FOR-SUPPORT-VECTOR-MACHINE/m-p/767310#M94750</guid>
      <dc:creator>Victor_G</dc:creator>
      <dc:date>2024-06-20T12:39:30Z</dc:date>
    </item>
    <item>
      <title>Re: SVEM ADD IN FOR SUPPORT VECTOR MACHINE</title>
      <link>https://community.jmp.com/t5/Discussions/SVEM-ADD-IN-FOR-SUPPORT-VECTOR-MACHINE/m-p/919861#M107885</link>
      <description>&lt;P&gt;Hi Victor, I tried to use this for mixture results, got a nice model but optimization with the profiler did not work because the component su did not add up to 1, was almost 3.. The same&amp;nbsp; happens in other ML models, PLS and even Gen Reg&lt;/P&gt;</description>
      <pubDate>Tue, 16 Dec 2025 14:50:12 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/SVEM-ADD-IN-FOR-SUPPORT-VECTOR-MACHINE/m-p/919861#M107885</guid>
      <dc:creator>frankderuyck</dc:creator>
      <dc:date>2025-12-16T14:50:12Z</dc:date>
    </item>
    <item>
      <title>Re: SVEM ADD IN FOR SUPPORT VECTOR MACHINE</title>
      <link>https://community.jmp.com/t5/Discussions/SVEM-ADD-IN-FOR-SUPPORT-VECTOR-MACHINE/m-p/919885#M107889</link>
      <description>&lt;P&gt;Hi Franck,&lt;/P&gt;
&lt;P&gt;Have you checked that your mixture factors have the right column properties :&amp;nbsp;&lt;/P&gt;
&lt;UL&gt;
&lt;LI&gt;"Mixture" with low and high individual factors values&lt;/LI&gt;
&lt;LI&gt;"Design Role" with role set as Mixture&lt;/LI&gt;
&lt;LI&gt;"Factor Changes" set at Easy (or other type if you have a split-plot/restricted randomization scenario)&lt;/LI&gt;
&lt;/UL&gt;
&lt;P&gt;Here is an example on Donev Mixture Data with factor CuSO4 :&lt;/P&gt;
&lt;P&gt;&lt;span class="lia-inline-image-display-wrapper lia-image-align-inline" image-alt="Victor_G_0-1765901003253.png" style="width: 400px;"&gt;&lt;img src="https://community.jmp.com/t5/image/serverpage/image-id/89045i3BD5859A1354D312/image-size/medium?v=v2&amp;amp;px=400" role="button" title="Victor_G_0-1765901003253.png" alt="Victor_G_0-1765901003253.png" /&gt;&lt;/span&gt;&lt;/P&gt;
&lt;P&gt;When launching the Profiler from Gaussian Process (or any other ML model), the mixture constraint is respected and enforced (mixture factors always sum up to 1):&lt;/P&gt;
&lt;P&gt;&lt;div class="lia-vid-container video-embed-center"&gt;&lt;div id="lia-vid-6386561719112w778h338r31" class="lia-video-brightcove-player-container"&gt;&lt;video-js data-video-id="6386561719112" data-account="6058004218001" data-player="default" data-embed="default" class="vjs-fluid" controls="" data-application-id="" style="width: 100%; height: 100%;"&gt;&lt;/video-js&gt;&lt;/div&gt;&lt;script src="https://players.brightcove.net/6058004218001/default_default/index.min.js"&gt;&lt;/script&gt;&lt;script&gt;(function() {  var wrapper = document.getElementById('lia-vid-6386561719112w778h338r31');  var videoEl = wrapper ? wrapper.querySelector('video-js') : null;  if (videoEl) {     if (window.videojs) {       window.videojs(videoEl).ready(function() {         this.on('loadedmetadata', function() {           this.el().querySelectorAll('.vjs-load-progress div[data-start]').forEach(function(bar) {             bar.setAttribute('role', 'presentation');             bar.setAttribute('aria-hidden', 'true');           });         });       });     }  }})();&lt;/script&gt;&lt;a class="video-embed-link" href="https://community.jmp.com/t5/video/gallerypage/video-id/6386561719112"&gt;(view in My Videos)&lt;/a&gt;&lt;/div&gt;&lt;/P&gt;
&lt;P&gt;Hope this answer may solve your problem,&lt;/P&gt;</description>
      <pubDate>Tue, 16 Dec 2025 16:06:16 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/SVEM-ADD-IN-FOR-SUPPORT-VECTOR-MACHINE/m-p/919885#M107889</guid>
      <dc:creator>Victor_G</dc:creator>
      <dc:date>2025-12-16T16:06:16Z</dc:date>
    </item>
    <item>
      <title>Re: SVEM ADD IN FOR SUPPORT VECTOR MACHINE</title>
      <link>https://community.jmp.com/t5/Discussions/SVEM-ADD-IN-FOR-SUPPORT-VECTOR-MACHINE/m-p/920276#M107928</link>
      <description>&lt;P&gt;OK Victor, thanks!&lt;/P&gt;</description>
      <pubDate>Thu, 18 Dec 2025 06:37:38 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/SVEM-ADD-IN-FOR-SUPPORT-VECTOR-MACHINE/m-p/920276#M107928</guid>
      <dc:creator>frankderuyck</dc:creator>
      <dc:date>2025-12-18T06:37:38Z</dc:date>
    </item>
  </channel>
</rss>

