I am using the Simple Probit Analysis script add-in to determine LD50, LD90, and LD95.
Is this script capable of correcting with Abbott's, or how do you set this up? Currently, I corrected the data myself by applying the Abbott's correction to the raw data to adjust for mortalities.
After running the Probit Add-In, how do I determine goodness of fit? There are no Chi Squared results. Is there another way I can run this analysis to get my LD values, as well as getting the Chi Square?
Here is an example of the data:
| Dose (% Conc.) | Start (#live) | Mortality @ 15 min. (#dead) |
| 100.00 | 13 | 13 |
| 100.00 | 10 | 10 |
| 100.00 | 11 | 11 |
| 100.00 | 13 | 13 |
| 25.00 | 10 | 9 |
| 25.00 | 23 | 17 |
| 25.00 | 15 | 9 |
| 25.00 | 9 | 6 |
| 10.00 | 14 | 3 |
| 10.00 | 10 | 1 |
| 10.00 | 10 | 4 |
| 10.00 | 16 | 7 |
| 1.64 | 11 | 1 |
| 1.64 | 12 | 3 |
| 1.64 | 13 | 4 |
| 1.64 | 24 | 4 |
I am trying to get LD50/LD90/LD95 results for dosage recommendations. I have this information for 5, 10, and 15 minute applications.
This is the data for my control:
| Dose (% Conc.) | Start (#live) | Mortality (#dead) |
| 0 | 10 | 1 |
| 0 | 11 | 6 |
| 0 | 10 | 3 |
| 0 | 15 | 4 |
OK, I think I understand now. I was correct to begin with and it is not a survival analysis (life time). We are back to my GLM example but now I will use your data. Here is your data in a JMP data table:
Set up the Fit Model launch dialog as before:
Click Run and then you get these results:
Click the red triangle next to Generalized Linear Model Fit and select Inverse Prediction. Type 0.5 (and any others that you want) for Probability (Died) and click OK.
So now we agree, no?
What do I do when some of the models have a bad fit?
Here is one of the data sets in particular I'm having issues with. If I Hide/Exclude the 0s, the model will fit.
You didn't answer any of my questions.
This example must be a different agent than the example that you tested at 15 minutes. It doesn't make sense that the same agent would be that much more lethal with just five more minutes of exposure.
You must use a different dilution/dose range (higher dilution, lower dose) to observe the transition. The highest dose should be 1.64. You should include the zero in the test and in the data analysis. This new data is the only way to fit the probit model for this agent at this time point.
I cannot speak to the discrepancies that you cite, but they call the data into question.
The problem with fitting is that the data exhibits the 'separation' problem. At one level (dose = 0), you have the low response and at all of the other levels (dose > 0), you have the high response. The transition from low to high response is impossible to determine. Try to find LD50 on your own with this data and you will understand the issue. (Does it occur at 0.1? At 0.5? At 1.5?)
You must perform another dose-response test where the dose ranges from 0 to 1.64 in order to fit the probit model and estimate LD50.
Excluding the high responses is not acceptable from a modeling or statistics point of view. You are suggesting that you exclude valid data to overcome a basic flaw with the study. (I understand that you could not foresee the potency of this agent, but that does not remove the flaw in the data.) You need more data, between dose = 0 and dose = 1.64, not less data, to solve this problem.
Excluding the high dose samples might be acceptable in your situation but I could not justify it. The fact that you (possibly) obtain a 'better' model this way is not validation. You simply cannot determine the LD50 from this sample of dose-response data except to say that LD50 occurs between 0 and 1.64.