SaraCHenriques ☆ Portugal, 20200701 14:37 (293 d 12:53 ago) Posting: # 21631 Views: 3,140 

Hi everyone, What is your opinion on below quantification limit (BQL) substitution for concentration data summary statistics, by timepoint? Zero substitution is the one I have seen the most, however, I don't think it is suitable for the calculation of geometric means. Could you give me your opinion on this matter? Many thanks! — Sara 
Helmut ★★★ Vienna, Austria, 20200701 15:27 (293 d 12:03 ago) @ SaraCHenriques Posting: # 21632 Views: 2,554 

Hi Sara, » What is your opinion on below quantification limit (BQL) substitution for concentration data summary statistics, by timepoint? You can use the median and quartiles or \(\small{\bar{x}_{geo}\mp SD_{geo}}\) if a certain percentage of samples are measurable (I have seen SOPs with 50%, 67%, and 75%) and nothing (‘not reportable’) otherwise. At the end of the day it’s not important at all (not relevant for the BE assessment). Use whatever you like.^{ 1} See also this (lengthy) thread. » Zero substitution is the one I have seen the most, … To quote Harold Boxenbaum (Crystal City workshop about bioanalytical method validation, Arlington 1990): After a dose we know only one thing for sure: The concentration is not zero. Ended in shouting matches.» … I don't think it is suitable for the calculation of geometric means. Correct, since$$\lim_{x \to 0} \log x=\infty.$$For simplicity we can say that \(\small{\log 0}\) is undefined. It is reasonable to assume that concentrations (\(\small{x \in \mathbb{R}^+}\)) follow a lognormal distribution, and the geometric mean would be the best estimator of location. Some people chicken out, set BQLs to zero, and present arithmetic means. This leads to funny plots with \(\small{\bar{x}\mp SD,}\) where the lower whisker reaches far below zero.^{ 2} Phew, negative concentrations? Not in this universe.
— Diftor heh smusma 🖖 Helmut Schütz The quality of responses received is directly proportional to the quality of the question asked. 🚮 Science Quotes 
martin ★★ Austria, 20200702 09:37 (292 d 17:53 ago) @ Helmut Posting: # 21634 Views: 2,434 

Dear Helmut and Sara, In statistical language  values <LLOQ are informatively censored and ignoring this fact can lead to biased results. Of course, adequate handling would require some modeling which seems to be in contradiction to regulatory thinking as NCA is clearly favored in BA/BE studies. However, I would like to bring a recent paper on this topic to your attention: Barnett H, Geys H, Jacobs T, Jaki T (2020). Methods for NonCompartmental Pharmacokinetic Analysis with Observations below the Limit of Quantification. Statistics in Biopharmaceutical Research, 123 best regards & hope this helps Martin 
Helmut ★★★ Vienna, Austria, 20200702 11:15 (292 d 16:15 ago) @ martin Posting: # 21635 Views: 2,432 

Hi Martin, » […] adequate handling would require some modeling which seems to be in contradiction to regulatory thinking as NCA is clearly favored in BA/BE studies. Yep. The EMA’s BEGL states: Noncompartmental methods should be used for determination of pharmacokinetic parameters in bioequivalence studies. The use of compartmental methods for the estimation of parameters is not acceptable. » However, I would like to bring a recent paper on this topic to your attention: […] Why am I not surprised that Thomas recommends kernel density imputation? However, PK modeling is not applied. \(\small{\widehat{AUC}}\) is still obtained by NCA, right? — Diftor heh smusma 🖖 Helmut Schütz The quality of responses received is directly proportional to the quality of the question asked. 🚮 Science Quotes 
ElMaestro ★★★ Denmark, 20200702 13:48 (292 d 13:42 ago) @ martin Posting: # 21636 Views: 2,374 

Hi Martin, » In statistical language  values <LLOQ are informatively censored and ignoring this fact can lead to biased results. » » Of course, adequate handling would require some modeling which seems to be in contradiction to regulatory thinking as NCA is clearly favored in BA/BE studies. However, I would like to bring a recent paper on this topic to your attention: Barnett H, Geys H, Jacobs T, Jaki T (2020). Methods for NonCompartmental Pharmacokinetic Analysis with Observations below the Limit of Quantification. Statistics in Biopharmaceutical Research, 123 Thanks for the reference. I'd love to see a work where someone not only debates BLQ's but also where someone tries to discuss what the various options imply for the residual variability and thus confidence interval in BE trials; the same for missing values. I have a feeling it might not be a big deal, but I dare not say at this point what "big deal" is quantitatively. — Pass or fail! ElMaestro 
martin ★★ Austria, 20200702 14:52 (292 d 12:38 ago) @ ElMaestro Posting: # 21637 Views: 2,346 

Dear ElMaestro, I am happy to hear that this information was considered as useful. Regarding missing values in BE trials: you may find this work of interest. Of note, this work addresses missing values from a conceptual rather than technical point of view such as the impact on the width of a CI. best regards & hope this helps Martin 
Ben ★ 20200707 16:37 (287 d 10:53 ago) @ martin Posting: # 21655 Views: 2,151 

Dear All » Of course, adequate handling would require some modeling which seems to be in contradiction to regulatory thinking as NCA is clearly favored in BA/BE studies. However, I would like to bring a recent paper on this topic to your attention: Barnett H, Geys H, Jacobs T, Jaki T (2020). Methods for NonCompartmental Pharmacokinetic Analysis with Observations below the Limit of Quantification. Statistics in Biopharmaceutical Research, 123 Well, another approach would be to fight for getting all measured values, including the ones where the lab tells us they are flagged as BLQ (yes, yes, I know they are "not reliable", whatever that means). But I have the feeling it is still better to use them than to ignore or set them to some fixed value. I had a nice discussion with Helmut about this topic. Any idea how this approach would behave? (could not see it within the article, question is also how to simulate such unreliable data... is it just data with higher variability?) Best regards, Ben. 
Helmut ★★★ Vienna, Austria, 20200708 10:23 (286 d 17:07 ago) @ Ben Posting: # 21656 Views: 2,147 

Hi Ben and all, » Well, another approach would be to fight for getting all measured values, including the ones where the lab tells us they are flagged as BLQ … Join “El ingenioso hidalgo Don Quijote de la Mancha” in his tilting at windmills… » … (yes, yes, I know they are "not reliable", whatever that means). Easy. At the LLOQ: Accuracy ≤20% and precision ≤20% in chromatography, ≤30% in ligand binding assays. Can be higher, if justified (hard data demonstrating that it is impossible to comply with the rules). <nitpick> In chemistry we are bound to the IUPAC’s terminology: inaccuracy, imprecision. » But I have the feeling it is still better to use them than to ignore or set them to some fixed value. Agree. Gutfeeling as well. » I had a nice discussion with Helmut about this topic. A summary: At the first Crystal City meeting about bioanalytical method validation (Arlington 1990) there were heated debates about the topic. Essentially there were two parties: Regulators wanted to have a ‘general rule’ in order to avoid discussions with applicants. Members of the PK modeling community strongly opposed that:
There is no reason not to use these values. It is just silliness that chemists fail to give you the measurements because of an arbitrary cut off that has no real meaning for pharmacokinetic analysis. Omitting these values will always cause bias. Hence, in my CRO we had an SOP:
» … question is also how to simulate such unreliable data... By (truncated?) lognormal distributions. » is it just data with higher variability? Nope. Higher (in)accuracy as well.
— Diftor heh smusma 🖖 Helmut Schütz The quality of responses received is directly proportional to the quality of the question asked. 🚮 Science Quotes 
ElMaestro ★★★ Denmark, 20200708 11:35 (286 d 15:55 ago) @ Ben Posting: # 21657 Views: 2,200 

Hi all, I like the discussion, but I wonder are BLQ's really such an issue? I mean, we have some rules which we can scientifically debate, but regardless of whether we like them scientifically, do the rules as we know them today lead to actual trouble? Just think about it, one perspective: Regulators mandate the use of the normal linear model in BE, and everyone knows that model may be right or wrong, probably it is wrong to a varying degree in all datasets. Sometimes we even know from KS tests or SW tests or God knows what, that the assumption of normality is outright wrong, yet we have to use the assumption of normality anyway. Is it a problem that an unknown proportion of studies definitely do not meet the assumptions, and that those that don't do so with a magnitude and nature whose consequences cannot be assessed? No, actually BE seems to work rather fine in spite of all this. At least as I see it. Somehow I see the BLQ discussion the same way. Yes, it may not be optimal, but hey it provides a strict and welldefined way forward where everyone can easily reproduce everything from raw data. Last time I looked people taking generics were not dropping dead in the streets A biased estimator may indeed be a useful estimator. A BLQ rule which is not scientifically optimal when looked at through the keyhole may still in a wider perspective be a good BLQ rule. How about the AUCinf extrapolation rule of 80%? Is it a complete disaster that it isn't 83%? I mean at the end of the day, the smaller the extrapolated area the better we know the profile from the getgo, so 83% must be better than 80%. And so forth. Just trying to put BE things into a perspective here. — Pass or fail! ElMaestro 
mittyri ★★ Russia, 20200708 12:03 (286 d 15:27 ago) @ Ben Posting: # 21658 Views: 2,075 

Dear Ben, I think you'll find that discussion useful where anchor points of Roger's position are given. I'd recommend to read Ohlbe's answer carefully. — Kind regards, Mittyri 