Acceptability of software [General Statistics]
Hi Lucas,
THX; reasonably good.
Good question! In the past I compared WinNonlin’s output obtained from datasets given in the literature (see here). I survived one inspection in 2004.
Software has to be validated, right? Unfortunately all these datasets were balanced crossovers. We recently published a paper comparing eight datasets evaluated by five different software packages – EquivTest/PK, Kinetica 5.0.10, SAS 9.2, Phoenix/WinNonlin 6.3.0.395, and R 3.0.2 (see here). We obtained identical results – with the exception of Kinetica, which screwed up all imbalanced datasets (PE & CI wrong). Thermo Fisher Scientific will correct the bug (see here).
Amazingly one of the reviewers of the manuscript asked “Was the software used in the analyses validated?” which we answered “For the purpose of bioequivalence the softwares were not validated beyond installation and where possible automatic IQ/OQ/PQ. The point of the paper is that it is virtually impossible to do in absence of a range of datasets, and the paper makes a proposal on which datasets such a validation could be based.”
Last Wednesday another paper dealing with 2-group parallel designs was accepted by the AAPS Journal. In PHX/WNL one needs a workaround to adjust for unequal variances (i.e., Welch Satterthwaite approximation) and the software is limited to 1,000 subjects/group (not an issue in BE). You can make an educated guess about how Kinetica performed.
We “validated” PHX/WNL’s templates for FDA’s RSABE and EMA’s ABEL (see here) by comparing its results with SAS’. Still some work to be done. EMA’s fully replicated dataset is imbalanced (nRTRT ≠ nTRTR) and incomplete (periods missing) whereas the partial replicate is both balanced and complete. We learned (from examples John posted here and others in Pharsight’s Extranet) that FDA’s ABE code for the partial replicate fails to converge sometimes due to the overspecified model (both in PHX/WNL and SAS). It seems that
❝ Hi Helmut, how are you?!
THX; reasonably good.

❝ Did you validated all workflow templates that you use on WinNonlin…
Good question! In the past I compared WinNonlin’s output obtained from datasets given in the literature (see here). I survived one inspection in 2004.
❝ …or there wasn't need for that?
Software has to be validated, right? Unfortunately all these datasets were balanced crossovers. We recently published a paper comparing eight datasets evaluated by five different software packages – EquivTest/PK, Kinetica 5.0.10, SAS 9.2, Phoenix/WinNonlin 6.3.0.395, and R 3.0.2 (see here). We obtained identical results – with the exception of Kinetica, which screwed up all imbalanced datasets (PE & CI wrong). Thermo Fisher Scientific will correct the bug (see here).
Amazingly one of the reviewers of the manuscript asked “Was the software used in the analyses validated?” which we answered “For the purpose of bioequivalence the softwares were not validated beyond installation and where possible automatic IQ/OQ/PQ. The point of the paper is that it is virtually impossible to do in absence of a range of datasets, and the paper makes a proposal on which datasets such a validation could be based.”
Last Wednesday another paper dealing with 2-group parallel designs was accepted by the AAPS Journal. In PHX/WNL one needs a workaround to adjust for unequal variances (i.e., Welch Satterthwaite approximation) and the software is limited to 1,000 subjects/group (not an issue in BE). You can make an educated guess about how Kinetica performed.
We “validated” PHX/WNL’s templates for FDA’s RSABE and EMA’s ABEL (see here) by comparing its results with SAS’. Still some work to be done. EMA’s fully replicated dataset is imbalanced (nRTRT ≠ nTRTR) and incomplete (periods missing) whereas the partial replicate is both balanced and complete. We learned (from examples John posted here and others in Pharsight’s Extranet) that FDA’s ABE code for the partial replicate fails to converge sometimes due to the overspecified model (both in PHX/WNL and SAS). It seems that
FA0(1)
instead of FA0(2)
could do the job. CSH
(suggested in FDA’s guidance) did not work for these nasty datasets.—
Dif-tor heh smusma 🖖🏼 Довге життя Україна!![[image]](https://static.bebac.at/pics/Blue_and_yellow_ribbon_UA.png)
Helmut Schütz
![[image]](https://static.bebac.at/img/CC by.png)
The quality of responses received is directly proportional to the quality of the question asked. 🚮
Science Quotes
Dif-tor heh smusma 🖖🏼 Довге життя Україна!
![[image]](https://static.bebac.at/pics/Blue_and_yellow_ribbon_UA.png)
Helmut Schütz
![[image]](https://static.bebac.at/img/CC by.png)
The quality of responses received is directly proportional to the quality of the question asked. 🚮
Science Quotes
Complete thread:
- SAS PROC GLM with group effect Silva 2014-05-19 10:57 [General Statistics]
- SAS PROC GLM with group effect ElMaestro 2014-05-19 11:51
- SAS PROC GLM with group effect Silva 2014-05-22 12:01
- SAS PROC GLM with group effect d_labes 2014-05-22 14:11
- SAS PROC GLM with group effect Silva 2014-05-22 16:54
- SAS PROC GLM with group effect d_labes 2014-05-22 14:11
- SAS PROC GLM with group effect Silva 2014-05-22 12:01
- SAS PROC GLM with group effect d_labes 2014-05-19 13:02
- SAS PROC GLM with group effect Silva 2014-05-23 12:12
- SAS PROC GLM with group effect ElMaestro 2014-05-23 12:31
- Acceptability of software Helmut 2014-05-23 12:47
- Acceptability of software Lucas 2014-11-14 12:21
- Acceptability of softwareHelmut 2014-11-14 14:11
- Acceptability of software Lucas 2014-11-14 12:21
- SAS PROC GLM with group effect ElMaestro 2014-05-19 11:51