<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Re: Split Plot DOE discussion time in Discussions</title>
    <link>https://community.jmp.com/t5/Discussions/Split-Plot-DOE-discussion-time/m-p/879008#M104225</link>
    <description>&lt;P&gt;Hi&amp;nbsp;&lt;a href="https://community.jmp.com/t5/user/viewprofilepage/user-id/3194"&gt;@Dan_Obermiller&lt;/a&gt;&amp;nbsp;,&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&amp;nbsp; I understand your scenario and your comments about the error distributions. As far as the responses of my DOE go, it's like I replied to&amp;nbsp;&lt;a href="https://community.jmp.com/t5/user/viewprofilepage/user-id/7073"&gt;@MRB3855&lt;/a&gt;&amp;nbsp;, I don't care what the distributions are from the standpoint of the response -- they could be bimodal, log-normal, SHASH, what have you. They are what they are for various (mostly known) reasons, and there is no indication a special cause variation took place during the DOE that would eliminate any of the data.&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&amp;nbsp; The problem is how to manage this during modeling/analysis as several different JMP Community and Help resources have conflicting recommendations. Some maintain it's critical to use the mixed model because it keeps the whole&amp;amp;random effects in the model, but also comes with it's own set of downsides (one is non-physical prediction values). On the other hand, GenReg can manage the non-normal responses and whole plot effects (and provide physically valid predictions), but it can't manage the random effects. The mixed model also results in normally distributed residuals with mean near 0, while GenReg does not. What's a good approach and why? (I'm not saying the "right" approach, but a good approach). I have also tried analyzing the data using the GLMM platform, which does an OK job for 3 of the 4 responses, but fails very badly with the one data set that is log-normal and bounded by values &amp;gt;=0.&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&amp;nbsp; As this is not an ideal case, where most of the JMP sample data tend to be, what are some best practices for managing the analysis when it's not ideal? The same goes for the other general questions I posed.&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;DS&lt;/P&gt;</description>
    <pubDate>Mon, 09 Jun 2025 15:43:25 GMT</pubDate>
    <dc:creator>SDF1</dc:creator>
    <dc:date>2025-06-09T15:43:25Z</dc:date>
    <item>
      <title>Split Plot DOE discussion time</title>
      <link>https://community.jmp.com/t5/Discussions/Split-Plot-DOE-discussion-time/m-p/878744#M104171</link>
      <description>&lt;P&gt;Hello JMP Community,&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&amp;nbsp; Running JMP Pro 18.1.1 on Windows 11. Apologies ahead of time for the long post, but I want to try and provide as much information ahead of time. The references to factor and response columns are for the attached anonymized data table for those interested in looking in more detail (hence some vagueness in my description and anonymized data table).&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&amp;nbsp; I have some general concerns and questions to bring up regarding a split-plot DOE that I'm helping my colleagues to analyze. Some not too dissimilar from the recent post &lt;A href="https://community.jmp.com/t5/Discussions/When-conducting-analysis-of-split-plot-design-experiments-if-the/m-p/878531#M104151" target="_self"&gt;here&lt;/A&gt;. Unlike the "clean" and "nice" examples in the JMP help data, this is a real-life industrial example where things don't always work out in an ideal way.&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&lt;U&gt;&lt;STRONG&gt;Background:&lt;/STRONG&gt;&lt;/U&gt; I have some chemist colleagues that wanted to run a DOE by testing out six factors and measuring four different responses. The purpose of the DOE was to hopefully determine what combination and at what levels those six factors could/should be set in order to achieve the desired results.&amp;nbsp; One of the responses is, what I would call for lack of better words, the "primary" response, Y__1, (the other responses have dependencies on this "primary response"). The dependency is NOT linear for each of the other responses.&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&lt;STRONG&gt;&lt;U&gt;DOE design:&lt;/U&gt;&lt;/STRONG&gt; We have a 30-run (this number was determined as the allowable number of runs by the chemists) custom DOE with six factors, one of which is hard to control (oven temperature), hence turning what would otherwise be a random design into a split-plot design.&lt;/P&gt;
&lt;UL&gt;
&lt;LI&gt;X__5 is the hard to change oven temperature response.&lt;/LI&gt;
&lt;LI&gt;Runs are grouped into 4 whole pots with 7 or 8 runs within each whole plot.&lt;/LI&gt;
&lt;LI&gt;Three of the four responses have lower design limits, with a goal to maximize their response (Y__1, Y__2, Y__3).&lt;/LI&gt;
&lt;LI&gt;One response has an upper design limit, with a goal to minimize its response (Y__4).&lt;/LI&gt;
&lt;LI&gt;One response has a lower detection limit of 10 and absolute lower limit of 0 (Y__3).&amp;nbsp;&lt;/LI&gt;
&lt;LI&gt;One response has an upper detection limit of 280 (Y__4).&lt;/LI&gt;
&lt;/UL&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&lt;U&gt;&lt;STRONG&gt;Observations of data:&lt;/STRONG&gt;&lt;/U&gt; The DOE was conducted and here are some observations of the data.&lt;/P&gt;
&lt;UL&gt;
&lt;LI&gt;Y__1 is normally distributed. Recall this is also the "primary" response.&lt;/LI&gt;
&lt;LI&gt;Y__2 is NOT normally distributed, but is best fit by a SHASH distribution. JMP does have a SHASH transform function to turn it into a normally distributed data set as well as an inverse SHASH transformation. Y__2 is not normal because it strongly depends on Y__1, and as Y__1 changes (decreases), Y__2 increases, goes through a maximum and then begins to decrease.&lt;/LI&gt;
&lt;LI&gt;Y__3 is NOT normally distributed because of the physical limit of 0 and detection limit of 10. This data is best characterized with a log-normal distribution. If using that as the Distribution in the GenReg platform, then the 0s need to be recoded as something very small and close to, but not 0, like 1e-12 or something.&lt;/LI&gt;
&lt;LI&gt;Y__4 is NOT normally distributed and is also better characterized by a log-normal distribution (using the updated continuous fit functions that can deal with detection limits vs the legacy fitters that suggest a normal 2-mixture instead).&lt;/LI&gt;
&lt;/UL&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&lt;U&gt;&lt;STRONG&gt;Discussion points/things to consider:&lt;/STRONG&gt;&lt;/U&gt;&amp;nbsp;Below are just some topics and discussions points that have either been brought up to me about the DOE results, or that I'm wondering how to manage/work with as I analyze the data.&lt;/P&gt;
&lt;UL&gt;
&lt;LI&gt;As was mentioned in the post linked above, because this is a split-plot DOE, it's important to keep the whole plot &amp;amp; random effects in the model because it's not a truly random DOE. Removing it can potentially lead to erroneous conclusions.
&lt;UL&gt;
&lt;LI&gt;Other platforms like Generalized Regression and Generalized Linear Mixed Models can't handle the random effects, but it can manage the whole plot as a factor.
&lt;UL&gt;
&lt;LI&gt;Does this mean I should stay away from using those platforms as a way to generate alternative models?&lt;/LI&gt;
&lt;/UL&gt;
&lt;/LI&gt;
&lt;LI&gt;Unfortunately, the only modeling platform that can handle the problem of detection limits for two of my responses is GenReg by using a Censor column. The others can't manage this, nor can they manage the non-normal distributions of 3 of my responses.
&lt;UL&gt;
&lt;LI&gt;NOTE: the censor column for Y__3 is different from Y__4 because the detection limits are different.&lt;/LI&gt;
&lt;/UL&gt;
&lt;/LI&gt;
&lt;LI&gt;This results in a dilemma: use the Mixed Model (or SLS) approach that handles the whole &amp;amp; random effects, but can't manage the non-normally distributed data or the detection limit problem; or use an alternative platform like GenReg which can handle the censored data and the detection limit problem, but can't handle the whole &amp;amp; random effects of the split-plot design.&lt;/LI&gt;
&lt;/UL&gt;
&lt;/LI&gt;
&lt;LI&gt;When running either the standard least squares or mixed model platforms with the data (or even GenReg and GLMM), all model profilers suggest that in order to maximize Y__1, Y__2, Y__3 and minimize Y__4, the factor settings should be set to an extreme value (either low or high -- it changes depending on the platform). This doesn't make sense from a domain knowledge perspective. Based on the ranges chosen for the DOE, we anticipated some/or all of the factors to be between the extremes for optimal responses.
&lt;UL&gt;
&lt;LI&gt;At present each response is being treated equally (25%, 0.25) in the set/maximize desirability options. This can change, but doesn't have a large effect on the profiler outcome -- it still suggests setting the factors at extreme values most of the time.&lt;/LI&gt;
&lt;/UL&gt;
&lt;/LI&gt;
&lt;LI&gt;The chemists have proposed augmenting the DOE by adding center points. I am not sure this would solve the problems we are facing in the analysis. But anyway,&amp;nbsp;
&lt;UL&gt;
&lt;LI&gt;I have tried to see if this is a possibility, but when I try this, I have to include the whole plots as a factor, and my only Augmentation Choice is "Augment", all the others are grayed out.
&lt;UL&gt;
&lt;LI&gt;Why do I not have the other options available?&lt;/LI&gt;
&lt;LI&gt;Is there a way to add center points or do a space filling without adding some kind of bias to the DOE?&lt;/LI&gt;
&lt;/UL&gt;
&lt;/LI&gt;
&lt;/UL&gt;
&lt;/LI&gt;
&lt;/UL&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&lt;U&gt;&lt;STRONG&gt;Questions/issues I'm struggling with:&lt;/STRONG&gt;&lt;/U&gt;&amp;nbsp;Overall, here are the issues I'm trying to manage when analyzing the data -- particularly in trying to figure out and understand how to avoid the profiler from suggesting extreme settings for the factors.&lt;/P&gt;
&lt;UL&gt;
&lt;LI&gt;Should I stick with the Mixed Model platform because of the split-plot design no matter what?&lt;/LI&gt;
&lt;LI&gt;Is it at all useful to try and use other platforms (GenReg, GLMM, etc.) to try and analyze the data?
&lt;UL&gt;
&lt;LI&gt;If it is worthwhile, how best to try and include the whole plot &amp;amp; random effects inherent in the DOE?&lt;/LI&gt;
&lt;/UL&gt;
&lt;/LI&gt;
&lt;LI&gt;What are some best practice methods for managing the strong dependence Y__2, Y__3, and Y__4 have on Y__1 in the analysis?&lt;/LI&gt;
&lt;LI&gt;What are some best practice methods for managing the non-linear responses of Y__2, Y__3, and Y__4? Transforming them, or is there some other/better way?&lt;/LI&gt;
&lt;LI&gt;I'd prefer to get a model for all responses at once, like how SLS can do it, but I don't think that platform is the correct platform to use given the non-normal distributions as well as dependence of the other responses on Y__1. I can save the prediction formulas for the responses from whatever platform I use and then use the Graph &amp;gt; Profiler to generate a prediction profiler, but I'm still stuck with the profiler suggesting extreme settings for the factors, which doesn't make sense.
&lt;UL&gt;
&lt;LI&gt;Why is the profiler suggesting extreme values?&lt;/LI&gt;
&lt;LI&gt;Does this indicate a problem with the DOE? If so, what is/are the problem(s).&lt;/LI&gt;
&lt;LI&gt;Can the extreme profiler suggestions be resolved somehow?&lt;/LI&gt;
&lt;/UL&gt;
&lt;/LI&gt;
&lt;LI&gt;If augmenting the design by performing more experiments is one way to go, how can I access other augmentation options? Right now, all I can do is change the upper/lower factor settings and define the number of additional runs, I can't do anything else.&lt;/LI&gt;
&lt;/UL&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&amp;nbsp; Thank you for taking the time to read through this post. Any feedback/thoughts/suggestions are much appreciated.&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;Thanks!,&lt;/P&gt;
&lt;P&gt;DS&lt;/P&gt;</description>
      <pubDate>Fri, 06 Jun 2025 19:16:33 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/Split-Plot-DOE-discussion-time/m-p/878744#M104171</guid>
      <dc:creator>SDF1</dc:creator>
      <dc:date>2025-06-06T19:16:33Z</dc:date>
    </item>
    <item>
      <title>Re: Split Plot DOE discussion time</title>
      <link>https://community.jmp.com/t5/Discussions/Split-Plot-DOE-discussion-time/m-p/878822#M104183</link>
      <description>&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;how are you ,&lt;SPAN&gt;Y__4 because&amp;nbsp;&amp;nbsp;has an upper design limit ??/&lt;/SPAN&gt;&lt;/P&gt;</description>
      <pubDate>Sun, 08 Jun 2025 04:27:13 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/Split-Plot-DOE-discussion-time/m-p/878822#M104183</guid>
      <dc:creator>MarkovHedgehog9</dc:creator>
      <dc:date>2025-06-08T04:27:13Z</dc:date>
    </item>
    <item>
      <title>Re: Split Plot DOE discussion time</title>
      <link>https://community.jmp.com/t5/Discussions/Split-Plot-DOE-discussion-time/m-p/878866#M104194</link>
      <description>&lt;P&gt;Your discussion is way more than a discussion about split-plots (response variables have nothing to do with split-plots). &amp;nbsp;Did you read the paper re. split-plots I suggested in the discussion you linked?&lt;/P&gt;</description>
      <pubDate>Sun, 08 Jun 2025 14:33:20 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/Split-Plot-DOE-discussion-time/m-p/878866#M104194</guid>
      <dc:creator>statman</dc:creator>
      <dc:date>2025-06-08T14:33:20Z</dc:date>
    </item>
    <item>
      <title>Re: Split Plot DOE discussion time</title>
      <link>https://community.jmp.com/t5/Discussions/Split-Plot-DOE-discussion-time/m-p/878898#M104202</link>
      <description>&lt;P&gt;&lt;a href="https://community.jmp.com/t5/user/viewprofilepage/user-id/4358"&gt;@statman&lt;/a&gt;&amp;nbsp;is correct that this is way more than a discussion. One thing that I will point out is that the distribution of your response variables is irrelevant. Your response data should have signals in it, which would alter the distribution.&amp;nbsp; The fact that some responses have a non-normal distribution should not be a surprise. The residuals (after the model is fit and signals are explained) are what need to be normally distributed.&lt;/P&gt;</description>
      <pubDate>Sun, 08 Jun 2025 23:41:27 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/Split-Plot-DOE-discussion-time/m-p/878898#M104202</guid>
      <dc:creator>Dan_Obermiller</dc:creator>
      <dc:date>2025-06-08T23:41:27Z</dc:date>
    </item>
    <item>
      <title>Re: Split Plot DOE discussion time</title>
      <link>https://community.jmp.com/t5/Discussions/Split-Plot-DOE-discussion-time/m-p/878967#M104210</link>
      <description>&lt;P&gt;I don't understand your question.&lt;/P&gt;</description>
      <pubDate>Mon, 09 Jun 2025 11:52:02 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/Split-Plot-DOE-discussion-time/m-p/878967#M104210</guid>
      <dc:creator>SDF1</dc:creator>
      <dc:date>2025-06-09T11:52:02Z</dc:date>
    </item>
    <item>
      <title>Re: Split Plot DOE discussion time</title>
      <link>https://community.jmp.com/t5/Discussions/Split-Plot-DOE-discussion-time/m-p/878968#M104211</link>
      <description>&lt;P&gt;I don't have easy access to the article, so I have not been able to read it.&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;I am reading this article, which has been somewhat helpful: Split-Plot Designs: What, Why, and How, B. Jones and C. J. Nachtsheim, J. Quality Technology, V. 41, N. 4, p. 340.&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Mon, 09 Jun 2025 11:56:25 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/Split-Plot-DOE-discussion-time/m-p/878968#M104211</guid>
      <dc:creator>SDF1</dc:creator>
      <dc:date>2025-06-09T11:56:25Z</dc:date>
    </item>
    <item>
      <title>Re: Split Plot DOE discussion time</title>
      <link>https://community.jmp.com/t5/Discussions/Split-Plot-DOE-discussion-time/m-p/878982#M104215</link>
      <description>&lt;P&gt;It may be a complicated and detailed discussion, but still a discussion nonetheless.&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;Semantics aside, I find it interesting that you mention the distribution of the response is irrelevant -- this seems to be counter to what was discussed by other JMP colleagues during a Discover Summit back in 2021 or 2022. In the presentation, there was a discussion about the use of the new detection limit column property for JMP 16. In the presentation, there was quite some commentary about the non-normal distribution of the response and how this needed to be properly accounted for by using a log-normal distribution when modeling the data with the GenReg platform. So, what is the correct approach -- do we treat the non-normal distribution of the response as irrelevant, or do I take it into consideration? It's hard to see how both can be right. I completely get your point about the residuals being normally distributed.&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;As for the other questions I posed, are there any thoughts on those? For example, the issue that the profiler us suggesting only extreme values as optimal settings for the combination of responses that we have. From our domain knowledge, this doesn't make sense. Is this an indication something has gone wrong with the DOE or some other systematic noise has been introduced into the data? Or, is this some kind of other indication that something else is not working properly.&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;Although I pose some specific questions to my case at hand, I also pose some very general questions that I would imagine others have also come across (or might come across in the future), hence their general relevance to the community. Many of the questions are generally about how to manage and handle the data when the analysis is not an ideal situation. Most, if not all of the example data provided with JMP tend to show ideal situations, but reality tends to be messier and needs more of a nuanced approach.&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;I'm still hopeful to have continued discussions, especially about the more general questions.&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;Thanks,&lt;/P&gt;
&lt;P&gt;DS&lt;/P&gt;</description>
      <pubDate>Mon, 09 Jun 2025 12:54:59 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/Split-Plot-DOE-discussion-time/m-p/878982#M104215</guid>
      <dc:creator>SDF1</dc:creator>
      <dc:date>2025-06-09T12:54:59Z</dc:date>
    </item>
    <item>
      <title>Re: Split Plot DOE discussion time</title>
      <link>https://community.jmp.com/t5/Discussions/Split-Plot-DOE-discussion-time/m-p/878999#M104219</link>
      <description>&lt;P&gt;Hi&amp;nbsp;&lt;a href="https://community.jmp.com/t5/user/viewprofilepage/user-id/12549"&gt;@SDF1&lt;/a&gt;&amp;nbsp;:I think we need some clarity around exactly what is meant when we talk about the normality of the response (we'll call it Y for now). And, to simplify,&amp;nbsp; let's just consider the OLS multiple regression model (Ordinary Least Squares). We can talk about normality in at least two different ways (for brevity, I'll leave out the subscripts and iid assumption):&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp;1. Y = a + b*x1 + c*x2 + d*x3 + e, where e is normally distributed with mean=0, and variance = Sigma^2.&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;Equivalently we could say&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp;2. Y is normally distributed with mean&amp;nbsp; a + b*x1 + c*x2 + d*x3, and variance = sigma^2.&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;And discussions about the distribution of the response is implicitly speaking to the way the model is described in (2) above. But, that is very different than saying, for example, that your Y column should have a normal distribution that can be assessed via the distribution platform (it has to be corrected for a mean that varies with the X's first), i.e., whichever way the model is described, the way we &lt;EM&gt;assess&lt;/EM&gt; normality is the same (via exploring the residuals).&amp;nbsp; So the normality (or not) of the response matters (in fact, parameter estimation in your mixed model via Fit Model/REML assumes normality), and it shouldn't be confused with how we assess normality.&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Mon, 09 Jun 2025 13:43:00 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/Split-Plot-DOE-discussion-time/m-p/878999#M104219</guid>
      <dc:creator>MRB3855</dc:creator>
      <dc:date>2025-06-09T13:43:00Z</dc:date>
    </item>
    <item>
      <title>Re: Split Plot DOE discussion time</title>
      <link>https://community.jmp.com/t5/Discussions/Split-Plot-DOE-discussion-time/m-p/879001#M104220</link>
      <description>&lt;P&gt;Let me try to clarify my comment. Saying the distribution of the response is irrelevant is a bit too strong, really. But let me paint a picture.&lt;/P&gt;
&lt;P&gt;Suppose I am measuring the blood pressure of some patients. All of these patients have high blood pressure. After taking the measurement, I give them all a very powerful blood pressure medication. After a month, I measure their blood pressure again. Would you expect the distribution of all blood pressures to be normally distributed? The answer is no. The medication would (hopefully) make the distribution bimodal. The signal is in the raw data and needs to be removed before checking for normality. All of the statistical inferences are based on the errors or residuals being normally distributed.&lt;/P&gt;
&lt;P&gt;Now as far as what other JMP colleagues were saying, I cannot comment. I was not there. However, if experience suggests that the variable has a certain distribution in a steady state, then you should use that certain distribution. That is commonly what is done when specifying an error distribution. Beyond that, I would need to have been part of that conversation to comment more.&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Mon, 09 Jun 2025 13:50:53 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/Split-Plot-DOE-discussion-time/m-p/879001#M104220</guid>
      <dc:creator>Dan_Obermiller</dc:creator>
      <dc:date>2025-06-09T13:50:53Z</dc:date>
    </item>
    <item>
      <title>Re: Split Plot DOE discussion time</title>
      <link>https://community.jmp.com/t5/Discussions/Split-Plot-DOE-discussion-time/m-p/879005#M104222</link>
      <description>&lt;P&gt;Hi&amp;nbsp;&lt;a href="https://community.jmp.com/t5/user/viewprofilepage/user-id/7073"&gt;@MRB3855&lt;/a&gt;&amp;nbsp;,&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&amp;nbsp; Thanks for your response, and I completely agree with you. I will try to be a little more clear about my questions regarding normality of the responses.&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&amp;nbsp; I actually don't care that the responses aren't normal -- they are what they are for several reasons that are known and not surprising. I don't believe I indicated that those responses which are non-normal&amp;nbsp;&lt;EM&gt;should&lt;/EM&gt; be normal. If I did, my apologies, as I don't think they&amp;nbsp;&lt;EM&gt;should&lt;/EM&gt; be anything. They are what they are, and there is currently no indication that any special cause would remove any data from the DOE.&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&amp;nbsp; Y__2 is non-normal because it is highly depended (in a nonlinear way) on Y__1. Y__3 is non-normal in part because physically it cannot have a value &amp;lt;0, and due to the design of the DOE, there are some measurements that are near this physical limit (it's also highly dependent on Y__1). While Y__4 on the other hand is in part non-linear because it is also highly&amp;nbsp; dependent on Y__1 and has an upper detection limit of 280 -- after that, the results are maxed-out, and because of the design there are some runs where the response is maxed-out. This is all well-and-good. So far, to me, the DOE has done exactly what it was designed to do.&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&amp;nbsp; My questions/concerns regarding the normality of the responses is more about how to manage/deal with the non-normal response in the modeling step. Some resources in the JMP Community &amp;amp; Help that I've read seem to have conflicting recommendations, especially when it comes to a split-plot designs where at least one factor is hard to change. Some of the resources recommend that the analysis should keep the whole&amp;amp;random effects in there (hence you must use the mixed model method -- but you can't account for non-normal distributions in your response -- or the detection limit), and some suggest using the GenReg platform where you can account for the non-normal distributions (and detection limits), but you loose the random effects in the model.&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&amp;nbsp; There are obviously trade-offs to the different approaches, and I'm more concerned in minimizing how those trade-offs affect the analysis of the DOE -- when it comes to the normality of the responses. If I use the GenReg platform, I can't account for the random effects inherent in the split-plot design and therefore susceptible to both Type 1 and Type 2 errors -- I could conclude something is there when it's not, or conclude something is not there when it is. On the other hand, the mixed model can't handle the non-normal response distributions (but can handle the whole&amp;amp;random effects), which can sometimes lead to predictions that aren't physically real -- like values &amp;lt;0, which don't make sense.&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&amp;nbsp; So, how does one handle this in a real example? The sample data tables in JMP are nice and ideal, providing a clear analytical path, but they don't really address gray areas like this where a clear analytical path isn't so straightforward.&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&amp;nbsp; Ultimately, it would be helpful to have a model that has low error and that the residuals are normally distributed, and centered around 0. The mixed model platform results give this (but also give non-physical response predictions), whereas the GenReg platform does not, but does provide physically valid response predictions. Again, how does one handle/manage this in a real, non-ideal situation?&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&amp;nbsp; Keep in mind that I have several other more broad questions/concerns as well.&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&amp;nbsp; I hope this has helped to clear up some points, but also to direct discussion to the other more general questions like augmentation, or what are best practice approaches to managing analysis when your data doesn't fit nicely into one model or the other, or why does the profiler always suggest extreme settings for optimal responses, and is this an indication something larger is wrong?&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;DS&lt;/P&gt;</description>
      <pubDate>Mon, 09 Jun 2025 14:47:34 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/Split-Plot-DOE-discussion-time/m-p/879005#M104222</guid>
      <dc:creator>SDF1</dc:creator>
      <dc:date>2025-06-09T14:47:34Z</dc:date>
    </item>
    <item>
      <title>Re: Split Plot DOE discussion time</title>
      <link>https://community.jmp.com/t5/Discussions/Split-Plot-DOE-discussion-time/m-p/879008#M104225</link>
      <description>&lt;P&gt;Hi&amp;nbsp;&lt;a href="https://community.jmp.com/t5/user/viewprofilepage/user-id/3194"&gt;@Dan_Obermiller&lt;/a&gt;&amp;nbsp;,&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&amp;nbsp; I understand your scenario and your comments about the error distributions. As far as the responses of my DOE go, it's like I replied to&amp;nbsp;&lt;a href="https://community.jmp.com/t5/user/viewprofilepage/user-id/7073"&gt;@MRB3855&lt;/a&gt;&amp;nbsp;, I don't care what the distributions are from the standpoint of the response -- they could be bimodal, log-normal, SHASH, what have you. They are what they are for various (mostly known) reasons, and there is no indication a special cause variation took place during the DOE that would eliminate any of the data.&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&amp;nbsp; The problem is how to manage this during modeling/analysis as several different JMP Community and Help resources have conflicting recommendations. Some maintain it's critical to use the mixed model because it keeps the whole&amp;amp;random effects in the model, but also comes with it's own set of downsides (one is non-physical prediction values). On the other hand, GenReg can manage the non-normal responses and whole plot effects (and provide physically valid predictions), but it can't manage the random effects. The mixed model also results in normally distributed residuals with mean near 0, while GenReg does not. What's a good approach and why? (I'm not saying the "right" approach, but a good approach). I have also tried analyzing the data using the GLMM platform, which does an OK job for 3 of the 4 responses, but fails very badly with the one data set that is log-normal and bounded by values &amp;gt;=0.&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&amp;nbsp; As this is not an ideal case, where most of the JMP sample data tend to be, what are some best practices for managing the analysis when it's not ideal? The same goes for the other general questions I posed.&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;DS&lt;/P&gt;</description>
      <pubDate>Mon, 09 Jun 2025 15:43:25 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/Split-Plot-DOE-discussion-time/m-p/879008#M104225</guid>
      <dc:creator>SDF1</dc:creator>
      <dc:date>2025-06-09T15:43:25Z</dc:date>
    </item>
    <item>
      <title>Re: Split Plot DOE discussion time</title>
      <link>https://community.jmp.com/t5/Discussions/Split-Plot-DOE-discussion-time/m-p/879188#M104243</link>
      <description>&lt;P&gt;Hi&amp;nbsp;&lt;a href="https://community.jmp.com/t5/user/viewprofilepage/user-id/12549"&gt;@SDF1&lt;/a&gt;,&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;Tricky question and complex situation, but you're right, this is often the case, unlike "perfect" toy/learning datasets.&lt;/P&gt;
&lt;P&gt;Having faced similar (and not ideal) situations, here is what I could recommend from a practical point of view :&lt;BR /&gt;&lt;BR /&gt;&lt;/P&gt;
&lt;OL&gt;
&lt;LI&gt;&lt;EM&gt;&lt;STRONG&gt;Identify patterns and anomalies&lt;/STRONG&gt;&lt;/EM&gt; : Plot the data to check if you have strange patterns of the responses depending on your whole plot effect and if you can already identify relationships between your factors and responses (+ correlations between responses). You can use the platform Multivariate to identify correlations as well as potential outliers using the Outlier analysis available in this platform.&amp;nbsp;&lt;/LI&gt;
&lt;LI&gt;&lt;EM&gt;&lt;STRONG&gt;First model iteration&lt;/STRONG&gt;&lt;/EM&gt; :
&lt;UL&gt;
&lt;LI&gt;Proceed with the full assumed model. Make sure how to deal with the Whole plot effect : Random effect (influence on response variance) or Fixed effect (influence on mean response) ? This answer should be dictated by the "physical meaning" of this whole plot situation and what it represent. In both cases, fitting the full assumed model will help you to understand how much variability is captured through this effect (either importance on response mean or response variance depending on the type of effects). &lt;BR /&gt;On your datatable and considering your whole plot effect as random, you have various influence of this random effect depending on the response : from insignificant (for Y2), to medium (around 19 and 23% of the total variance for Y3 and Y1) to high (more than 40% of the variance captured by whole plot effect for Y4). Depending on this influence, this can give you an idea about how much precision you can lose by dropping this random effect out of the model. I would try to keep a mixed model for Y4, Y3 and Y1 if possible, and no need for mixed model for Y2.&lt;/LI&gt;
&lt;LI&gt;At this stage, you can check the residuals from your model (and Actual by Predicted plot), to see if you have curved pattern in your residual that may indicates a non-adequate model (missing term ?) or that a transformation may be needed.&amp;nbsp;&lt;BR /&gt;Be careful about transformation vs. generalized model, as they do not handle the data in the same way :&amp;nbsp;&lt;LI-MESSAGE title="Difference between &amp;amp;quot;least square&amp;amp;quot; and &amp;amp;quot;generelized linear method&amp;amp;quot; in the fit model" uid="638087" url="https://community.jmp.com/t5/Discussions/Difference-between-quot-least-square-quot-and-quot-generelized/m-p/638087#U638087" discussion_style_icon_css="lia-mention-container-editor-message lia-img-icon-forum-thread lia-fa-icon lia-fa-forum lia-fa-thread lia-fa"&gt;&lt;/LI-MESSAGE&gt;&amp;nbsp;Transformation affect your data directly (so the mean and variance of the response), whereas generalized models&amp;nbsp;&lt;SPAN&gt;use a link function to transform the mean into a linear function of the predictor variables and a variance function to allow for variance heterogeneity in the analysis rather than trying to transform it away. I wouldn't tranform the responses unless I have a strong indication that it may be needed (simplification of the model with transformed responses and better residual patterns by transforming the response, which seems to be the case for Y2).&lt;/SPAN&gt;&lt;/LI&gt;
&lt;LI&gt;&lt;SPAN&gt;Finally, you can also check which effects are statistically and practically significant, to understand in other models iterations which effects are interesting to consider including. You can also compare the results with other platforms, like&amp;nbsp;&lt;A href="https://www.jmp.com/support/help/en/18.2/#page/jmp/the-fit-two-level-screening-platform.shtml?_gl=1*ax9r37*_up*MQ..*_ga*MTE1Mjc1MjY5LjE3NDk1NjA3NzI.*_ga_BRNVBEC1RS*czE3NDk1NjA3NzEkbzEkZzAkdDE3NDk1NjA3NzEkajYwJGwwJGgw#" target="_blank" rel="noopener"&gt;The Fit Two Level Screening Platform&lt;/A&gt;&amp;nbsp;that can help identify active effects, or Fit Least Squares modeling with random effect.&lt;/SPAN&gt;&lt;/LI&gt;
&lt;/UL&gt;
&lt;/LI&gt;
&lt;LI&gt;&lt;SPAN&gt;&lt;EM&gt;&lt;STRONG&gt;Models exploration/comparison/selection&lt;/STRONG&gt;&lt;/EM&gt; : Use various platforms depending on the assumptions and previous results :&lt;/SPAN&gt;
&lt;UL&gt;
&lt;LI&gt;&lt;SPAN&gt;&lt;SPAN&gt;For response with medium/high impact of whole plot (random) effect on the response (Y4, Y3 and Y1), you can use Mixed Model or Least Squares model with whole plot random effect. To deal with censored data, you can use Weights in the model : create a numeric continuous column where censored data has a low value (binary weights : 0,5 for example, or based on ratio of censored on non-censored data, or more complex weights trend : the farther from the threshold limit you are, the lowest weight/importance your observation will have on model fitting) and non-censored data has a high value (binary weights 2 for example, or based on ratio of non-censored data on censored data, or more complex as described above). This will bias the model fitting and put more importance on data that is non-censored, and give little importance ("weight") to censored data (since it can't be exactly measured, so it is not critical to have good fit and low residuals in this censored area). You can then check the "Actual by Predicted" plot, see if the residuals seem acceptable for the non-censored measurement part, as well as checking residuals and verifying regression assumption for the non-censored data part.&lt;BR /&gt;Example with the response Y_3 with the weight formula using the inverse of the squared distance to the detection limit of 10:&amp;nbsp;&lt;/SPAN&gt;&lt;/SPAN&gt;
&lt;PRE&gt;&lt;CODE class=" language-jsl"&gt;If( :Y__3 &amp;gt; 10,
	1,
	1 / ((10 - :Y__3) ^ 2)
)&lt;/CODE&gt;&lt;/PRE&gt;
&lt;SPAN&gt;&lt;SPAN&gt;And the modeling results (with the whole plot random effect):&lt;BR /&gt;&lt;/SPAN&gt;&lt;/SPAN&gt;&lt;span class="lia-inline-image-display-wrapper lia-image-align-inline" image-alt="Victor_G_1-1749569223777.png" style="width: 400px;"&gt;&lt;img src="https://community.jmp.com/t5/image/serverpage/image-id/76754iE5A1C304F23DA6A1/image-size/medium?v=v2&amp;amp;px=400" role="button" title="Victor_G_1-1749569223777.png" alt="Victor_G_1-1749569223777.png" /&gt;&lt;/span&gt;&lt;SPAN&gt;&lt;BR /&gt;&lt;/SPAN&gt;&lt;/LI&gt;
&lt;LI&gt;&lt;SPAN&gt;For response with low impact&amp;nbsp;of whole plot (random) effect on the response (Y2), you can try different models with Fit Least Squares, GenReg for example. You can then compare effects included in the models, and compare models based on various criteria : statistical significance, RMSE, R²/R²-adjusted, AICc/BIC, ...&lt;/SPAN&gt;&lt;/LI&gt;
&lt;/UL&gt;
&lt;/LI&gt;
&lt;LI&gt;&lt;SPAN&gt;&lt;EM&gt;&lt;STRONG&gt;Validation &amp;amp; Augmentation&lt;/STRONG&gt;&lt;/EM&gt; : Validate your models using validation runs. It may seem strange that&amp;nbsp;extreme settings are suggested for the factors, but since your DoE mostly use only 2-levels factors, there is not much curvature (besides those brought by interaction effects) that could be present in your responses. &lt;BR /&gt;Augmenting your DoE to include 2nd order (or more) polynomial terms may be a good idea (instead of centre points that won't be helpful to estimate polynomial terms possibly contributing to curvature), to be able to draw a design space satisfying your 4 responses.&lt;/SPAN&gt;&lt;/LI&gt;
&lt;/OL&gt;
&lt;P&gt;&lt;SPAN&gt;Some other considerations :&lt;/SPAN&gt;&lt;/P&gt;
&lt;UL&gt;
&lt;LI&gt;&lt;SPAN&gt;Depending on when Y2, Y3 and Y4 are measured vs. Y1 (primary response of interest) and the relevance of this idea, you could maybe add these intermediate responses as predictors for Y1. The process could be then to predict Y2, Y3 and Y4 based on factors, and then predict Y1 based on other predicted responses + factors. There is a risk to inflate prediction error with this intermediate step, but it may be worth a try. You could also model Y1 thanks to Y2, Y3 and Y4 only, and using PLS or other modeling platforms able to deal with collinearity.&lt;/SPAN&gt;&lt;/LI&gt;
&lt;LI&gt;&lt;SPAN&gt;&lt;SPAN&gt;You could also try Machine Learning approaches, as they may be good comparative model to estimate factors importance, compared to "traditional" statistical modeling. Also, depending on how you would like the prediction profile to be, the "step-based" prediction profile of tree-based models may be interesting for censored data, to not predict precisely below a certain threshold for example.&lt;BR /&gt;Example here for Y4 with a simple partition tree with only 2 splits, you can see that X5 and X6 have the highest influence on Y4, and you can separate censored data from the rest of the data (with the second split at X6 &amp;lt; 65) :&lt;BR /&gt;&lt;/SPAN&gt;&lt;/SPAN&gt;&lt;span class="lia-inline-image-display-wrapper lia-image-align-inline" image-alt="Victor_G_0-1749568379724.png" style="width: 400px;"&gt;&lt;img src="https://community.jmp.com/t5/image/serverpage/image-id/76753i0053E50662E61D51/image-size/medium?v=v2&amp;amp;px=400" role="button" title="Victor_G_0-1749568379724.png" alt="Victor_G_0-1749568379724.png" /&gt;&lt;/span&gt;
&lt;P&gt;Even if this model may not be highly predictive or "practically convenient", this kind of analysis is helpful to compare with other models, and try to identify important predictors.&lt;/P&gt;
&lt;/LI&gt;
&lt;/UL&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;And at the end, a model isn't meant to be perfect, it's meant to be useful. There might be compromise in the modeling platforms chosen, but you should be able to justify your choices based on validation data and domain expertise.&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;Hope this (long) answer might provide some useful ideas,&lt;/P&gt;</description>
      <pubDate>Tue, 10 Jun 2025 16:03:40 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/Split-Plot-DOE-discussion-time/m-p/879188#M104243</guid>
      <dc:creator>Victor_G</dc:creator>
      <dc:date>2025-06-10T16:03:40Z</dc:date>
    </item>
    <item>
      <title>Re: Split Plot DOE discussion time</title>
      <link>https://community.jmp.com/t5/Discussions/Split-Plot-DOE-discussion-time/m-p/879190#M104245</link>
      <description>&lt;P&gt;Victor I agree completely with your first step, but, I would like to propose an alternate to your step #2. &amp;nbsp;You should &lt;STRONG&gt;not&lt;/STRONG&gt; fit a full model as there are 2 different error structures (whole plot and sub plot). &amp;nbsp;This would make for inappropriate comparisons. &amp;nbsp;For example, comparing the WP factor(s) to the MSE of the subplot for statistical significance is comparing apples to oranges. &amp;nbsp;That p-value is useless. You should essentially treat the whole plot and sub plot as if they are 2 different experiments. &amp;nbsp;Please read Box and Jones. &amp;nbsp;Also Anderson and McLean, Sanders, Bisgaard.&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P style="font-weight: 400;"&gt;Box, G.E.P., Stephen Jones (1992), “&lt;EM&gt;Split-plot designs for robust product experimentation&lt;/EM&gt;”, &lt;U&gt;Journal of Applied Statistics&lt;/U&gt;, Vol. 19, No. 1&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P style="font-weight: 400;"&gt;Jones, Bradley, Christopher J. Nachtsheim (2009) &lt;EM&gt;“Split-Plot Designs: What, Why, and How”,&lt;/EM&gt; &lt;U&gt;Journal of Quality Technology&lt;/U&gt;, Vol. 41, No. 4, pp. 340-361&lt;/P&gt;
&lt;P style="font-weight: 400;"&gt;&amp;nbsp;&lt;/P&gt;
&lt;P style="font-weight: 400;"&gt;Anderson, Virgil and McLean, Robert (1974) “&lt;EM&gt;Design of Experiments, A Realistic Approach&lt;/EM&gt;” Marcell Decker (ISBN 0-8247-7493-0)&lt;/P&gt;
&lt;P style="font-weight: 400;"&gt;&amp;nbsp;&lt;/P&gt;
&lt;P style="font-weight: 400;"&gt;Sanders, D., &amp;amp; Coleman, J. (2003). Recognition and Importance of Restrictions on Randomization in Industrial Experimentation. &lt;I&gt;Quality Engineering&lt;/I&gt;, &lt;I&gt;15&lt;/I&gt;(4), 533–543. &lt;A href="https://doi.org/10.1081/QEN-120018386" target="_blank" rel="noopener"&gt;https://doi.org/10.1081/QEN-120018386&lt;/A&gt;&lt;/P&gt;
&lt;P style="font-weight: 400;"&gt;&amp;nbsp;&lt;/P&gt;
&lt;P style="font-weight: 400;"&gt;&lt;SPAN&gt;Bisgaard, S. (2000). The Design and Analysis of 2&lt;/SPAN&gt;&lt;SUP&gt;&lt;I&gt;k–p&lt;/I&gt;&lt;/SUP&gt;&lt;SPAN&gt; × 2&lt;/SPAN&gt;&lt;SUP&gt;&lt;I&gt;q–r&lt;/I&gt;&lt;/SUP&gt;&lt;SPAN&gt; Split Plot Experiments. &lt;/SPAN&gt;&lt;I&gt;Journal of Quality Technology&lt;/I&gt;&lt;SPAN&gt;, &lt;/SPAN&gt;&lt;I&gt;32&lt;/I&gt;&lt;SPAN&gt;(1), 39–56. &lt;A href="https://doi.org/10.1080/00224065.2000.11979970" target="_blank" rel="noopener"&gt;https://doi.org/10.1080/00224065.2000.11979970&lt;/A&gt;&lt;/SPAN&gt;&lt;/P&gt;
&lt;P style="font-weight: 400;"&gt;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Tue, 10 Jun 2025 19:40:12 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/Split-Plot-DOE-discussion-time/m-p/879190#M104245</guid>
      <dc:creator>statman</dc:creator>
      <dc:date>2025-06-10T19:40:12Z</dc:date>
    </item>
    <item>
      <title>Re: Split Plot DOE discussion time</title>
      <link>https://community.jmp.com/t5/Discussions/Split-Plot-DOE-discussion-time/m-p/879934#M104345</link>
      <description>&lt;P&gt;Personal Perspective on Split-Plot Experimental Design:‌&lt;BR /&gt;Before conducting the formal experimental design, fix all factors at their normal production levels and collect data by blocking. Analyze the significance of the random block effect. If the random block effect is not significant, proceed with conventional experimental design methods for the formal experiment, but impose randomization restrictions on some factors during execution and use standard analytical methods. If the random block effect is significant, adopt a split-plot design.&lt;/P&gt;</description>
      <pubDate>Sat, 14 Jun 2025 12:00:55 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/Split-Plot-DOE-discussion-time/m-p/879934#M104345</guid>
      <dc:creator>Rily_Maya</dc:creator>
      <dc:date>2025-06-14T12:00:55Z</dc:date>
    </item>
    <item>
      <title>Re: Split Plot DOE discussion time</title>
      <link>https://community.jmp.com/t5/Discussions/Split-Plot-DOE-discussion-time/m-p/880407#M104411</link>
      <description>&lt;P&gt;Hi&amp;nbsp;&lt;a href="https://community.jmp.com/t5/user/viewprofilepage/user-id/11568"&gt;@Victor_G&lt;/a&gt;&amp;nbsp;,&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&amp;nbsp; Thanks for taking the time to write your detailed response and provide some thoughts on the problems at hand. I've been working through some of the suggestions you provided as well as pursuing some other avenues and ideas.&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&amp;nbsp; In response to your thoughts, here is some feedback:&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;OL&gt;
&lt;LI&gt;&lt;STRONG&gt;&lt;EM&gt;Identify patterns and&lt;/EM&gt; anomalies&lt;/STRONG&gt;: Yes, I have looked into that, and one thing that I noticed here is that Whole Plots 2 &amp;amp; 3, where X__5 is at it's "low" setting results in generally overall poor performance for the 4 responses. This is not too surprising for us. It's good that we had this in the factor settings so we could see a large enough "signal" change in the responses. As mentioned before, 3 of the 4 responses are correlated (non-linearly) and dependent on the first response, Y__1, so seeing those correlations in the Multivariate platform are not new for us. One of the difficulties here is that for one of the responses, Y__2, it is non-monotonic as a function of Y__1 -- meaning that Y__2 can have the same response value at different Y__1 values. For example, Y__2 can have a value of 24 when Y__1 is either ~27 or ~31. Y__3 and Y__4 don't have this behavior, yet they are still non-linearly dependent on Y__1. There is one potential outlier based on the Y__3 response, but so far, we have no special cause reason to eliminate this run or to try and re-do it. However, it is something I will ask my colleagues to look further into.&lt;/LI&gt;
&lt;LI&gt;&lt;STRONG&gt;&lt;EM&gt;First model iteration&lt;/EM&gt;&lt;/STRONG&gt;:
&lt;UL&gt;
&lt;LI&gt;I agree that the Whole Plot effect is random -- ideally, the DOE cold have been done as completely randomized, but because X__5 is hard to change (at least when trying to do this DOE as efficiently as possible), it ended up being a split-plot design. Anyway, based on what we know about some of the responses and how they can be influenced by X__5, we can't really drop the Whole Plots effect. And, although Y__2 doesn't really need it to be included, it is easier to fit the entire Mixed Model as a whole.&lt;/LI&gt;
&lt;LI&gt;Thanks for the thoughts on the transformations. In testing out the SHASH transformation for Y__2, it actually did not fit as well as just fitting Y__2 without transformation. I have been evaluating the residuals as well as other model assessment methods like doing a Bland-Altman analysis or looking at the 95% CI for the actual by predicted plots. So far, the Mixed Model seems to be performing the best across the model assessment methods.&lt;/LI&gt;
&lt;/UL&gt;
&lt;/LI&gt;
&lt;LI&gt;&lt;EM&gt;&lt;STRONG&gt;Models/exploration/comparison/selection&lt;/STRONG&gt;&lt;/EM&gt;:&amp;nbsp;
&lt;UL&gt;
&lt;LI&gt;While keeping the random effect in the model, the Mixed Model still seems to do a little better job than the Least Squares. Both of them predict some slightly non-realistic values for one of the responses (a different one for each model). They're not drastically unrealistic, but given what we know, especially about response Y__1, those responses should not be able to be achieved.&lt;/LI&gt;
&lt;LI&gt;Thanks for the suggestion on the weighted column based on how far from the detection limit the response is. I like that and have adopted it when comparing different models while trying to account for the detection limit issue.&lt;/LI&gt;
&lt;LI&gt;Based on the censored/weighted responses as well as that some require the random effect and one doesn't, I might end up blending prediction columns from different platforms, but I need to further evaluate model performance.&lt;/LI&gt;
&lt;/UL&gt;
&lt;/LI&gt;
&lt;LI&gt;&lt;EM&gt;&lt;STRONG&gt;Validation &amp;amp; Augmentation&lt;/STRONG&gt;&lt;/EM&gt;&lt;EM&gt;:&lt;/EM&gt;
&lt;UL&gt;
&lt;LI&gt;Definitely will want to validate the model -- that's always what I recommend (I can't force them to do anything, just recommend).&lt;/LI&gt;
&lt;LI&gt;As far as the profiler goes, it's just strange that according to the optimization criteria the profiler always goes to extreme values, which seems a bit strange and unusual.&lt;/LI&gt;
&lt;LI&gt;As far as augmentation goes, I was hoping to just take the original DOE and do some space filling runs, but it won't let me do that because of the hard to change factor. Instead, if I take your approach and actually change the model by adding in quadratic effects, then I can have JMP generate runs that aren't just repeating the corners/edges of the design space and actually put runs inside the space, which I hope will help.&lt;/LI&gt;
&lt;/UL&gt;
&lt;/LI&gt;
&lt;/OL&gt;
&lt;P&gt;Your other considerations:&lt;/P&gt;
&lt;UL&gt;
&lt;LI&gt;Sorry, should have been a bit more clear on this one. All the responses can be measured independent of one another -- they could theoretically be done simultaneously (in parallel) or sequentially (in serial). The order of the testing is actually irrelevant. What I was trying to explain is that Y__2, Y__3, and Y__4 all depend on Y__1 and each in a non-linear way. It would be better to consider using the factors+Y__1 Prediction Formula to predict the other responses. But yes, there is the risk of inflating the prediction error.&lt;/LI&gt;
&lt;LI&gt;I had not thought of using the Partition platform to analyze the data. Not to take the results too seriously, I think it is helpful to look at the analysis and use that information to our advantage. Thanks! Some of the other methods don't work as well, but I think it might be in part because we don't have so many runs.&lt;/LI&gt;
&lt;/UL&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&amp;nbsp; I fully agree about models being useful and not perfect -- but that is the whole crux of what I'm trying to do with the data set I have -- generate a useful model, and when a prediction profiler only suggests extreme settings (and inconsistently based on SLS or Mixed Model approaches), I wanted to look for ideas on how to make the most learning I can from this data set.&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&amp;nbsp; Yes, your response was very helpful and provided a lot of ideas for how to evaluate things. Thank you. I appreciate the discussion and effort to engage in the data and topic!&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;Thanks!,&lt;/P&gt;
&lt;P&gt;DS&lt;/P&gt;</description>
      <pubDate>Wed, 18 Jun 2025 15:23:42 GMT</pubDate>
      <guid>https://community.jmp.com/t5/Discussions/Split-Plot-DOE-discussion-time/m-p/880407#M104411</guid>
      <dc:creator>SDF1</dc:creator>
      <dc:date>2025-06-18T15:23:42Z</dc:date>
    </item>
  </channel>
</rss>

