<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Re: deployment model in Discussions</title>
    <link>https://community.jmp.com/t5/Discussions/deployment-model/m-p/373493#M62313</link>
    <description>&lt;P&gt;Meanwhile I believe that I will not be able to fit a valid/reasonable model for the data.&lt;/P&gt;&lt;P&gt;My first approach was to generate validation column statified by Xs.&lt;/P&gt;&lt;P&gt;The Timeseries hint from&amp;nbsp;&lt;a href="https://community.jmp.com/t5/user/viewprofilepage/user-id/1701"&gt;@dale_lehman&lt;/a&gt;&amp;nbsp;was useful, so I looked also at the data in a timeseries manner,&lt;/P&gt;&lt;P&gt;realizing that with time the combination of the Xes change. And here it again looks to me like "New" is different. Yes, the Xes cover the same range, but that is not enough. I think, the combination is different in "New", that means, it is a kind of extrapolation. This would be not an issue for regression, but for methods like RF this counts (model is multidimensional and highly nonlinear, so difficult to look inside).&lt;/P&gt;&lt;P&gt;On the other side I was not able to fit a reasonable model with any other method (neural, Boosted Tree). It looks to me that we do not have the proper information in the X-columns to build a solid model.&lt;/P&gt;&lt;P&gt;So I would look into the physics of the process or whatever it is,&lt;/P&gt;&lt;UL&gt;&lt;LI&gt;how the Xes may/can change the Y,&amp;nbsp;&lt;/LI&gt;&lt;LI&gt;what the noise may be&lt;/LI&gt;&lt;LI&gt;and what makes the things change over time, and include it into the dataset.&lt;/LI&gt;&lt;LI&gt;and I'm also curious about the data structure, there are groups of rows with same date and same values for some parameters&lt;/LI&gt;&lt;/UL&gt;&lt;P&gt;What me surprised is that, when I made a validation column that is stratified by time, I got the same result (fair model for the first part, no model with the new data).&lt;/P&gt;&lt;P&gt;Good luck,&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
    <pubDate>Thu, 01 Apr 2021 11:58:01 GMT</pubDate>
    <dc:creator>Georg</dc:creator>
    <dc:date>2021-04-01T11:58:01Z</dc:date>
    <item>
      <title>deployment model</title>
      <link>https://community.jmp.com/t5/Discussions/deployment-model/m-p/372727#M62263</link>
      <description>&lt;P&gt;Hi&lt;/P&gt;&lt;P&gt;to everybody.&lt;/P&gt;&lt;P&gt;In attached 2 jmp files. The one named Training-validation-test was used to build a RF model with the following results.&lt;/P&gt;&lt;P&gt;The file contains one numeric response and 20 different predictors with one of them categorical, The predictors&amp;nbsp;are sensors values. &amp;nbsp;&amp;nbsp;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&lt;span class="lia-inline-image-display-wrapper lia-image-align-inline" image-alt="Capture.PNG" style="width: 551px;"&gt;&lt;img src="https://community.jmp.com/t5/image/serverpage/image-id/31710i8D5D3DF36DAB8BB4/image-size/large?v=v2&amp;amp;px=999" role="button" title="Capture.PNG" alt="Capture.PNG" /&gt;&lt;/span&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;As you can see the Bias is very good while the variance is slightly high (92-69=23%).&lt;/P&gt;&lt;P&gt;Then a new database with examples never seen by the model was used to simulate the deployment of algorithm on new data.&amp;nbsp;&lt;/P&gt;&lt;P&gt;With high surprise the result was really poor with&amp;nbsp;a R2 between the predicted&amp;nbsp;and the actual response close to zero. We were not able to understand this behavior. The R2=68% for test data after all doesn't&amp;nbsp;seems to be really&amp;nbsp;bad. So we were expecting to obtain&amp;nbsp;for sure a R2 larger than zero.&lt;/P&gt;&lt;P&gt;Please any help/comment is welcome.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Felice-Gian&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Fri, 09 Jun 2023 00:31:21 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/deployment-model/m-p/372727#M62263</guid>
      <dc:creator>FR60</dc:creator>
      <dc:date>2023-06-09T00:31:21Z</dc:date>
    </item>
    <item>
      <title>Re: deployment model</title>
      <link>https://community.jmp.com/t5/Discussions/deployment-model/m-p/372774#M62268</link>
      <description>&lt;P&gt;Virtually all of the X variables (with the exception of X12, X13, and X19) are significantly different, on average, between the two sets.&amp;nbsp; So, my guess would be that the new data is not simply another random sample from the same population, but comes from a different population.&amp;nbsp; It is not surprising that your model, which fit the original data fairly well, does so poorly on new data, when that new data comes from a different population.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;I suppose the question to ask is why it comes from a different population?&amp;nbsp; Apparently you thought the data was being drawn from the same population (or something close to that).&amp;nbsp; But must of these variables look so different between the two sets that such an assumption is clearly not correct.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;To see the differences, I concatenated the 2 tables and just used Fit Y by X, with Y being all of your X variables and X being the source table created when I concatenated them.&lt;/P&gt;</description>
      <pubDate>Tue, 30 Mar 2021 18:11:27 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/deployment-model/m-p/372774#M62268</guid>
      <dc:creator>dale_lehman</dc:creator>
      <dc:date>2021-03-30T18:11:27Z</dc:date>
    </item>
    <item>
      <title>Re: deployment model</title>
      <link>https://community.jmp.com/t5/Discussions/deployment-model/m-p/372828#M62275</link>
      <description>&lt;P&gt;I agree to&amp;nbsp;&lt;a href="https://community.jmp.com/t5/user/viewprofilepage/user-id/1701"&gt;@dale_lehman&lt;/a&gt;&amp;nbsp;, the new data is different.&lt;/P&gt;&lt;P&gt;I was able to fit a similar model, working also poorly on new data.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&lt;span class="lia-inline-image-display-wrapper lia-image-align-inline" image-alt="Georg_0-1617137935128.png" style="width: 400px;"&gt;&lt;img src="https://community.jmp.com/t5/image/serverpage/image-id/31723iEF1676A456C0F4F9/image-size/medium?v=v2&amp;amp;px=400" role="button" title="Georg_0-1617137935128.png" alt="Georg_0-1617137935128.png" /&gt;&lt;/span&gt;&lt;/P&gt;&lt;P&gt;The Scatterplot matrix showing some details, e.g. X5 using very small range of the original data, same with X11 ...&lt;/P&gt;&lt;P&gt;In both pictures the new data is selected. Maybe this helps you to find the reason why there is a difference.&lt;/P&gt;&lt;P&gt;&lt;span class="lia-inline-image-display-wrapper lia-image-align-inline" image-alt="Georg_1-1617138125434.png" style="width: 400px;"&gt;&lt;img src="https://community.jmp.com/t5/image/serverpage/image-id/31724i2D643CC64E545490/image-size/medium?v=v2&amp;amp;px=400" role="button" title="Georg_1-1617138125434.png" alt="Georg_1-1617138125434.png" /&gt;&lt;/span&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Tue, 30 Mar 2021 21:05:15 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/deployment-model/m-p/372828#M62275</guid>
      <dc:creator>Georg</dc:creator>
      <dc:date>2021-03-30T21:05:15Z</dc:date>
    </item>
    <item>
      <title>Re: deployment model</title>
      <link>https://community.jmp.com/t5/Discussions/deployment-model/m-p/372927#M62281</link>
      <description>&lt;P&gt;Hi Dale and Georg&lt;/P&gt;&lt;P&gt;thanks for your reply.&lt;/P&gt;&lt;P&gt;Really the differences between the two distributions for each predictors to me don't seems to be so different (see histograms in the attached picture. With bls we indicate the Training-Val-Test set and exp the set never seen by the model). Maybe the X20 shows a new peak in the never seen data that was absent before. But the remaining distributions are only less populated then previous ones.&lt;/P&gt;&lt;P&gt;If the predictor distribution of training&amp;nbsp;set is wider than test one, the model should have "visited" the phase space occupied by the predictor of test set. Correct? So why it should do a bad prediction? &amp;nbsp;&lt;/P&gt;&lt;P&gt;Anyway I confirmed (even though for us this is strange) what you say by taking&amp;nbsp;randomly 2000 instances (instead of the last 2000 instances in thenever seen by the random forest model and once trained it was able to predict correctly the response. So if a slight difference in terms of average/std make a big difference in terms of prediction performance how we can use these models&amp;nbsp;in production where the signals change continuously?&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Rgds.&amp;nbsp; Felice/Gian&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Wed, 31 Mar 2021 08:44:47 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/deployment-model/m-p/372927#M62281</guid>
      <dc:creator>FR60</dc:creator>
      <dc:date>2021-03-31T08:44:47Z</dc:date>
    </item>
    <item>
      <title>Re: deployment model</title>
      <link>https://community.jmp.com/t5/Discussions/deployment-model/m-p/372977#M62287</link>
      <description>&lt;P&gt;Now that I look more closely, I see that this is time series data.&amp;nbsp; From my initial views, it does not appear that the process has changed significantly.&amp;nbsp; However, I think the problem is that the train-validate-test paradigm is not really appropriate for time series data.&amp;nbsp; I'd suggest creating a distinct time period (e.g., the last month of the first data set) as the validation set and re-estimating the model.&amp;nbsp; My guess will be that the model performance on the validation and never seen before data should be more similar if you estimate the model that way.&amp;nbsp; I'm still a bit concerned about the noticeable differences between the X variables in the two data sets - it might require some more sophisticated time series modeling (such as modeling potential seasonality).&lt;/P&gt;</description>
      <pubDate>Wed, 31 Mar 2021 13:21:38 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/deployment-model/m-p/372977#M62287</guid>
      <dc:creator>dale_lehman</dc:creator>
      <dc:date>2021-03-31T13:21:38Z</dc:date>
    </item>
    <item>
      <title>Re: deployment model</title>
      <link>https://community.jmp.com/t5/Discussions/deployment-model/m-p/373493#M62313</link>
      <description>&lt;P&gt;Meanwhile I believe that I will not be able to fit a valid/reasonable model for the data.&lt;/P&gt;&lt;P&gt;My first approach was to generate validation column statified by Xs.&lt;/P&gt;&lt;P&gt;The Timeseries hint from&amp;nbsp;&lt;a href="https://community.jmp.com/t5/user/viewprofilepage/user-id/1701"&gt;@dale_lehman&lt;/a&gt;&amp;nbsp;was useful, so I looked also at the data in a timeseries manner,&lt;/P&gt;&lt;P&gt;realizing that with time the combination of the Xes change. And here it again looks to me like "New" is different. Yes, the Xes cover the same range, but that is not enough. I think, the combination is different in "New", that means, it is a kind of extrapolation. This would be not an issue for regression, but for methods like RF this counts (model is multidimensional and highly nonlinear, so difficult to look inside).&lt;/P&gt;&lt;P&gt;On the other side I was not able to fit a reasonable model with any other method (neural, Boosted Tree). It looks to me that we do not have the proper information in the X-columns to build a solid model.&lt;/P&gt;&lt;P&gt;So I would look into the physics of the process or whatever it is,&lt;/P&gt;&lt;UL&gt;&lt;LI&gt;how the Xes may/can change the Y,&amp;nbsp;&lt;/LI&gt;&lt;LI&gt;what the noise may be&lt;/LI&gt;&lt;LI&gt;and what makes the things change over time, and include it into the dataset.&lt;/LI&gt;&lt;LI&gt;and I'm also curious about the data structure, there are groups of rows with same date and same values for some parameters&lt;/LI&gt;&lt;/UL&gt;&lt;P&gt;What me surprised is that, when I made a validation column that is stratified by time, I got the same result (fair model for the first part, no model with the new data).&lt;/P&gt;&lt;P&gt;Good luck,&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Thu, 01 Apr 2021 11:58:01 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/deployment-model/m-p/373493#M62313</guid>
      <dc:creator>Georg</dc:creator>
      <dc:date>2021-04-01T11:58:01Z</dc:date>
    </item>
    <item>
      <title>Re: deployment model</title>
      <link>https://community.jmp.com/t5/Discussions/deployment-model/m-p/373539#M62321</link>
      <description>&lt;P&gt;I have a couple of suggestions.&amp;nbsp; First, I would try first differencing your response variable and then trying random forests again.&amp;nbsp; For me, that is the easiest approach to time series data - and it is possible that the data will be better behaved when looked at that way.&amp;nbsp; I would still manually create the validation set as the last time period rather than a random selection of rows.&amp;nbsp; The second idea is something you noticed - there are time groupings, such as multiple observations closely linked in time.&amp;nbsp; You might consider making the time sequence uniform by summarizing observations closely related in time.&amp;nbsp; Time series are better behaved if the time variable is uniformly measured, so that means your data will need to be summarized into equal spaced time intervals.&amp;nbsp; As you are suggesting, it is time to go back to thinking about exactly what is being measured and how it is being measured.&lt;/P&gt;</description>
      <pubDate>Thu, 01 Apr 2021 14:35:44 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/deployment-model/m-p/373539#M62321</guid>
      <dc:creator>dale_lehman</dc:creator>
      <dc:date>2021-04-01T14:35:44Z</dc:date>
    </item>
    <item>
      <title>Re: deployment model</title>
      <link>https://community.jmp.com/t5/Discussions/deployment-model/m-p/373943#M62362</link>
      <description>&lt;P&gt;Hi George and Dale&lt;/P&gt;&lt;P&gt;thank you very much for your answers. I give a look at the lag plot of all predictors and response and with my big surprise I discovered a very long autocorrelation (see attached pictures).&amp;nbsp;&lt;/P&gt;&lt;P&gt;Since the data come from a semiconductor fab, the value 25 visible in the ACF is related to the number of the wafers in a lot.&amp;nbsp;&amp;nbsp;&lt;/P&gt;&lt;P&gt;The points out line instead in the lag plots are just the first wafers of a different lot loaded in the macgine. I think that the autocorrelation is due to internal feedback present in our machine to keep the working point. Of course those adjustments are hidden in to the chamber and not under our control and understanding. The response also show a clear lag that I cannot explain. But is there.&amp;nbsp;&lt;/P&gt;&lt;P&gt;I tried to use the lag, moving average, smoothing and other predictors features engineering but the results continue to be really bad.&amp;nbsp;&lt;/P&gt;&lt;P&gt;Next step I want to try the differencing that Dale proposed. Do you believe that the simple difference X(t)-X(t-1) can be enough?&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Ciao Felice&amp;nbsp; &amp;nbsp;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Fri, 02 Apr 2021 15:34:24 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/deployment-model/m-p/373943#M62362</guid>
      <dc:creator>FR60</dc:creator>
      <dc:date>2021-04-02T15:34:24Z</dc:date>
    </item>
    <item>
      <title>Re: deployment model</title>
      <link>https://community.jmp.com/t5/Discussions/deployment-model/m-p/373976#M62366</link>
      <description>&lt;P&gt;The PACF plot suggests that the first two lags are the relevant ones - by far, the first lag is the most important.&amp;nbsp; So, I think simple first order differencing would be a good thing to try.&amp;nbsp; It may be possible that one of the machine learning models (e.g., RF, boosted trees, neural nets) will perform well, so I would try them.&amp;nbsp; But you also should try some more standard time series models (where you would not use the differencing, as the time series models are built to account for the autocorrelation in various ways).&amp;nbsp; In theory, if you use first order differencing (and if that largely eliminates the time structure), then you can probably use a random train-validate-test set for your modeling.&amp;nbsp; At least, that's what I think - others may know better.&lt;/P&gt;</description>
      <pubDate>Fri, 02 Apr 2021 18:04:58 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/deployment-model/m-p/373976#M62366</guid>
      <dc:creator>dale_lehman</dc:creator>
      <dc:date>2021-04-02T18:04:58Z</dc:date>
    </item>
    <item>
      <title>Re: deployment model</title>
      <link>https://community.jmp.com/t5/Discussions/deployment-model/m-p/374128#M62384</link>
      <description>&lt;P&gt;Hi Dale&amp;nbsp;&lt;/P&gt;&lt;P&gt;since also the predictors show the same PACF, do you think that could be useful differencing the predictors too beside the response?&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Thanks&amp;nbsp; Felice&lt;/P&gt;</description>
      <pubDate>Sat, 03 Apr 2021 09:36:45 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/deployment-model/m-p/374128#M62384</guid>
      <dc:creator>FR60</dc:creator>
      <dc:date>2021-04-03T09:36:45Z</dc:date>
    </item>
    <item>
      <title>Re: deployment model</title>
      <link>https://community.jmp.com/t5/Discussions/deployment-model/m-p/374129#M62385</link>
      <description>&lt;P&gt;Time series models are not my strength, so hopefully someone will give you a more definitive answer.&amp;nbsp; I would not difference the features at first - I'd build the models and look at the residuals to see if they exhibit the same autocorrelation.&amp;nbsp; If they do, then perhaps I'd try differencing some of the features.&lt;/P&gt;</description>
      <pubDate>Sat, 03 Apr 2021 11:03:09 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/deployment-model/m-p/374129#M62385</guid>
      <dc:creator>dale_lehman</dc:creator>
      <dc:date>2021-04-03T11:03:09Z</dc:date>
    </item>
    <item>
      <title>Re: deployment model</title>
      <link>https://community.jmp.com/t5/Discussions/deployment-model/m-p/374164#M62388</link>
      <description>&lt;P&gt;Hi Dale&amp;nbsp;&lt;/P&gt;&lt;P&gt;unfortunatly also by using the response variable difference none of the regeression models was able to return a good result.&amp;nbsp;&lt;/P&gt;&lt;P&gt;So at this point maybe I need to assume that our X variables are not good predictors of response.&amp;nbsp;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Anyway still remain the doubt if for data like ours (sensor data) we have to split the training/test by taking in account the time axis or instead&amp;nbsp; we can use the random split tecnique. In the literature I found several papers where for sensor data they use the random split training/test. By if we do in the same way the results seems really good until we do the deployment of model in to production to understand that the model is completely wrong. So where is the truth?&amp;nbsp;&lt;/P&gt;&lt;P&gt;In attached you can find a paper that apply the machine learning tecniques&amp;nbsp; to a semicondcutor tool (etcher). They try to do a predictive maintenance by looking at the sensors data. So the data should be really similar to ones. If you give a look at the part hof the paper ighlighted in green looks like that they split the data taking in account the time axis and then considering the data as time series.&amp;nbsp;&lt;/P&gt;&lt;P&gt;I hope that more people besides Dale can have their say and help us to solve this dilemma.&lt;/P&gt;&lt;P&gt;Are or not the data like the ours multi-time series?&amp;nbsp;&lt;/P&gt;&lt;P&gt;It is true that the predictors X and the respone Y depend on the time. But if we want predict the Y according to the X values why the time variable should be so important? To me the Y should depend on the X variables that are set automatically by the machine during the process phase. Of course this happen in the time cronologically, but the X value depend on the recipe and machine state. So I have some difficult to accept the dependancy on the time.&amp;nbsp;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Happy Easter to the whole community.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Felice&lt;/P&gt;</description>
      <pubDate>Sat, 03 Apr 2021 18:05:23 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/deployment-model/m-p/374164#M62388</guid>
      <dc:creator>FR60</dc:creator>
      <dc:date>2021-04-03T18:05:23Z</dc:date>
    </item>
    <item>
      <title>Re: deployment model</title>
      <link>https://community.jmp.com/t5/Discussions/deployment-model/m-p/374184#M62392</link>
      <description>&lt;P&gt;I strongly suspect you need to spend more time preparing the data.&amp;nbsp; To fit time series models, the observations must be equally spaced and with no duplicate time values.&amp;nbsp; You have 4 sensors (?) with unevenly spaced responses.&amp;nbsp; If you want to make sense of this data, I believe you need to aggregate the values so there is a single observation at each time period, and you might want to create a variable for the length of time between timed observations.&amp;nbsp; I'd also suggest doing a separate analysis for each type of sensor rather than using sensor as one of your features, as the behavior of each appears to be different.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;I wouldn't be prepared to conclude that your features are not predictive - I think too much noise is introduced by mixing the sensor types as well as having nonstandard time intervals.&amp;nbsp; If you address these issues, there may be a good predictive model yet.&amp;nbsp; The modeling part is not the problem - the data preparation is (as is often the case).&amp;nbsp; At least that is what I think.&lt;/P&gt;</description>
      <pubDate>Sat, 03 Apr 2021 20:47:36 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/deployment-model/m-p/374184#M62392</guid>
      <dc:creator>dale_lehman</dc:creator>
      <dc:date>2021-04-03T20:47:36Z</dc:date>
    </item>
    <item>
      <title>Re: deployment model</title>
      <link>https://community.jmp.com/t5/Discussions/deployment-model/m-p/375809#M62567</link>
      <description>&lt;P&gt;Mybe this can help in the discussion. I used the excellent model screening option in jmp pro by using the nested option to tune hyperparameters combined with the k-fold cross validation. Here the results obtained after some hour.of jmp work&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;span class="lia-inline-image-display-wrapper lia-image-align-inline" image-alt="Picture1.png" style="width: 604px;"&gt;&lt;img src="https://community.jmp.com/t5/image/serverpage/image-id/31979i239287D88E429E36/image-size/large?v=v2&amp;amp;px=999" role="button" title="Picture1.png" alt="Picture1.png" /&gt;&lt;/span&gt;&lt;/P&gt;&lt;P&gt;What was surprising me is the large difference among the models in terms of performances. The tree models and KNN showed a R2 above 50% while all the fit models showed values well below the 50% (close to zero) even though the lasso and ridge regularization usage. Mybe is this difference that explain the bad performance of RF/Boosted&amp;nbsp; models once deployed in production?&lt;/P&gt;&lt;P&gt;What do you think? Any comment/suggestion/help?&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Felice.&lt;/P&gt;</description>
      <pubDate>Sun, 11 Apr 2021 13:11:59 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/deployment-model/m-p/375809#M62567</guid>
      <dc:creator>FR60</dc:creator>
      <dc:date>2021-04-11T13:11:59Z</dc:date>
    </item>
  </channel>
</rss>

