Angie,
Sorry, I'm not sure I can help without understanding the situation. I thought you had sent fake data, but the situation was real...apologies. I have some further points regarding your replies (if you'll indulge me).
1. By "your alpha is .05" are you talking alpha risk? That is used for statistical significance. I am asking about practical significance. This is ALWAYS more important than statistical, because statistical significance is dependent on the data, how you acquired it, and what comparisons you make.
2. Not to be belligerent, but having the assay done in a lab, does not make it "good". I have many examples of lab measurements that were incapable of detecting that which they were intended to detect. Do you have data to support the precision of the assays? If so, by all means proceed.
3. The setting for the experiment is in the lab? What inference space do you hope to draw conclusions about? Will the conclusions be drawn on "conditions" outside of the lab? Is so, proceed with caution. You'll want to make sure the results can be applied to the area of concern. Again, my advice is limited as it depends on the actual situation.
OK, you have 3 "data points" for each treatment combination. The question is how were they gotten? (I'm being careful about what to call those 3 measures as different folks use different language to describe them). Some guidance, if the "data points" are acquired without the treatment combinations changing, they are typically not considered independent events and therefore do not increase the degrees of freedom in the study (I call them repeats). The reason they vary cannot be assigned to the treatments (factors) because those were "constant" when the 3 data points were taken, so the reason for those values varying is due to variables that change in a short time period, short-term noise (e.g., measurement error). You can, in this case, look at the variation within treatment (graphically works) and then determine the proper summary statistics (e.g., mean, variance, range...) which become the Y's to model.
If the 3 data points are indeed independent events, then they may be considered replicates, which increase the DF's available. Often the advice is to do randomized replicates as this provides, hopefully, an unbiased estimate of experimental error to be used in statistical tests. It also increases the inference space and perhaps provides some "unassigned" DF's to add covariates. In my opinion, there are other more efficient and effective ways to increase inference space, assign the effect of noise, partition the noise and thereby increase the precision of the design (detecting factor effects), create robust designs, etc. (RCBD, BIB, split-plots, etc.)
4. I'm not sure what you're saying here, but selection of response variables is critical. The more continuous the response variable, the more efficiently and effectively you can understand factor relationships. I always suggest to measure as much as is feasible, video the experiment, record sounds, etc. You never know when a new response variable leads to discovery. Certainly if the problem is one of variation, you should have a response variable that quantifies the variation (e.g., standard deviation, variance, range, etc.).
5. There is a host of advice regarding selection of the correct tool and number of samples needed. My advice is to start with a list of questions (e.g., What questions do you want to answer? What hypotheses, about dependent and independent variables, do you have? How will the phenomena be measured?, etc.). Once you understand the situation, what types of data are available, what effects you want to estimate, what resources are available and what is the sense of urgency, etc. you can design the study (sampling plan, experiment, etc.). I'm going to oversimplify...degrees of freedom is the amount of information you have in the data set. With the data set you sent me (where there was a time series), you had 15 data points so you have 14 total degrees of freedom. Each factor has multiple degrees of freedom ( e.g., Donor has 4, Tail Feather 4, Tale feather count 4) and then you have interactions (multiply the factor DF's). As you can imagine the number of degrees of freedom for the factors alone adds up to 12 degrees of freedom, so there are not enough DF's to estimate the interactions. The rule is you need enough DF's to estimate or compare what you want to estimate or compare. (sorry the # of DF's you need is completely dependent on the situation). I'll end with a paraphrase from one of my favorite DOE authors (Cuthbert Daniels):
The commonest of defects in DOE are:
- Oversaturation: too many effects for the number of treatments. (this is your issue)
- Overconservativeness: too many observations for the desired estimates
- Failure to study the data for bad values
- Failure to take into account all of the aliasing
- Imprecision due to misunderstanding the error variance.
That's all I got for now. Thanks for indulging me. Let me know if I can be of further assistance.
"All models are wrong, some are useful" G.E.P. Box