Correspondence Volume 122 | Issue 3 | March 2014
Environ Health Perspect; DOI:10.1289/ehp.1307727R
Instruments for Assessing Risk of Bias and Other Methodological Criteria: Krauth et al. Respond
David Krauth,1 Tracey J. Woodruff,2,3 Lisa Bero1,4
Citation: Krauth D, Woodruff TJ, Bero L. 2014. Instruments for assessing risk of bias and other methodological criteria: Krauth et al. respond. Environ Health Perspect 122:A67; http://dx.doi.org/10.1289/ehp.1307727R
The authors declare they have no actual or potential competing financial interests.
Published: 1 March 2014
Related EHP Correspondence
Beck et al. criticize our systematic review (Krauth et al. 2013) because we included instruments derived from preclinical animal research. Assessment instruments developed for preclinical animal models have criteria that are relevant to hazard and risk assessment because risk of bias in animal studies is not dependent on the data stream or the question being asked, but on the design of the study. Many instruments that have been developed (including those for evaluating animal toxicology studies) have criteria that have not been shown to bias research outcomes (see Supplemental Material, Table S1, of Krauth et al. 2013).
Furthermore, Table 1 of our paper (Krauth et al. 2013) lists the criteria found in most instruments we identified. In the “Discussion,” we described the empirical evidence supporting the use of some of these criteria and cited the relevant references with the empirical data. By empirical evidence, we mean that a criterion (e.g., randomization) has been shown to be associated with overestimation or underestimation of effect (this could be an efficacy or harm outcome).
Beck et al. note several publications in environmental chemical health hazard assessment [Ågerstrand et al. 2011; Food and Drug Administration (FDA) 2003; Hulzebos et al. 2010; Organisation for Economic Co-operation and Development (OECD) 1998; Schneider et al. 2009; U.S. Environmental Protection Agency (EPA) 1999a, 1999b, 2013]. All of these publications, except OECD (1998), were identified in our search; however, they did not meet the a priori inclusion criteria for our systematic review. As noted in our “Methods” (Krauth et al. 2013), we included the earliest publication of an instrument when it was used in subsequent reports. The article by Ågerstrand et al. (2011) was based on four earlier published papers (i.e., Durda and Preziosi 2000; Hobbs et al. 2005; Klimisch et al. 1997; Schneider et al. 2009). We cited three of these in our review, but excluded Schneider et al. (2009) because it appeared to be a description of software that could be used to operationalize the Klimisch criteria. After reviewing the criteria described by Schneider et al. (2009) in their supplemental file, we found no unique additional criteria that were not already included in our Table 1 and Supplemental Material, Table S1. The reports from the U.S. EPA (1999a, 1999b) and FDA (2003) were neither indexed in Medline nor found in screening of bibliographies. In addition, U.S. EPA (2013) was published after we ended our study. Because we did not find the OECD document (OECD 1998), we cannot conclude whether or not it should have been included in our study.
The comment by Beck et al. that the National Toxicology Program is relying on criteria that have not been “transparently empirically tested” is not correct. In our paper (Krauth et al. 2013), we recommended the use of empirically tested criteria and we pointed out criteria that have been shown to be a risk of bias.
We caution against gathering judgments on how to assess study quality and propose that evidence should guide such evaluations. We propose an empirically based approach—as opposed to consensus-based opinion of experts—as this would provide a more unbiased evaluation of the data.
Ågerstrand M, Breitholtz M, Ruden C. 2011. Comparison of four different methods for reliability evaluation of ecotoxicity data: a case study of non-standard test data used in environmental risk assessments of pharmaceutical substances. Environ Sci Eur 23:17; doi: 10.1186/2190-4715-23-17.
FDA (Food and Drug Administration). 2003. General Guidelines for Designing and Conducting Toxicity Studies. In: Guidance for Industry and Other Stakeholders, Toxicological Principles for the Safety Assessment of Food Ingredients, Redbook 2000. Available: http://www.fda.gov/Food/GuidanceRegulation/GuidanceDocumentsRegulatoryInformation/IngredientsAdditivesGRASPackaging/ucm078315.htm [accessed 15 October 2013].
Krauth D, Woodruff TJ, Bero L. 2013. Instruments for assessing risk of bias and other methodological criteria of published animal studies: a systematic review. Environ Health Perspect 121:985–992; doi: 10.1289/ehp.1206389.
OECD (Organisation for Economic Co-operation and Development). 1998. OECD Series on Principles of Good Laboratory Practice and Compliance Monitoring, No 1: OECD Principles on Good Laboratory Practice. ENV/MC/CHEM(98)17. Paris:OECD. Available: http://search.oecd.org/officialdocuments/displaydocumentpdf/?doclanguage=en&cote=env/mc/chem(98)17 [accessed 13 February 2014].
Schneider K, Schwarz M, Burkholder I, Kopp-Schneider A, Edler L, Kinsner-Ovaskainen A, et al. 2009. “ToxRTool”, a new tool to assess the reliability of toxicological data. Toxicol Lett 189(2):138–144.
U.S. EPA (U.S. Environmental Protection Agency). 1999a. Auditing General Toxicology Studies. Available: http://www.epa.gov/compliance/resources/policies/monitoring/fifra/sop/glp-da-09.pdf [accessed 15 October 2013].
U.S. EPA (U.S. Environmental Protection Agency). 1999b. Determining the Adequacy of Existing Data. Available: http://www.epa.gov/hpv/pubs/general/datadfin.htm [accessed 15 October 2013].
U.S. EPA (U.S. Environmental Protection Agency). 2013. OCSPP Harmonized Test Guidelines. Series 870: Health Effects Test Guidelines. Available: http://www.epa.gov/ocspp/pubs/frs/publications/Test_Guidelines/series870.htm [accessed 15 October 2013].
ISEE 2015 Abstracts Now AvailableEHP is pleased to present the abstracts for the 2015 annual conference of the International Society for Environmental Epidemiology (ISEE), “Addressing Environmental Health Inequalities,” held 30 August–3 September 2015 in São Paulo, Brazil.
CEHN September 2015 Article of the Month“Relation of Prenatal Methylmercury Exposure from Environmental Sources to Childhood IQ” (DOI:10.1289/ehp.1408554) has been selected by the Children’s Environmental Health Network (CEHN) as its September 2015 Article of the Month. These CEHN summaries discuss the potential policy implications of current children’s environmental health research.
New Editor-in-ChiefWe are pleased to announce that Sally Perreault Darney has been selected as the new Editor-in-Chief of EHP. Sally comes to EHP from the U.S. Environmental Protection Agency, where she most recently co-led a large research project focused on assessing health disparities in vulnerable groups and providing healthy environments for children. Learn more about Sally and her vision for the journal in the September issue of EHP.
Sign Up to Receive E-mail Alerts
Recent Advance Publications
Alternatives Assessment Frameworks: Research Needs for the Informed Substitution of Hazardous Chemicals
Occupation and Risk of Non-Hodgkin Lymphoma and Its Subtypes: A Pooled Analysis from the InterLymph Consortium
Blood Biomarkers of Late Pregnancy Exposure to Trihalomethanes in Drinking Water and Fetal Growth Measures and Gestational Age in a Chinese Cohort
Birth Weight, Ethnicity, and Exposure to Trihalomethanes and Haloacetic Acids in Drinking Water during Pregnancy in the Born in Bradford Cohort
Prenatal Ambient Air Pollution, Placental Mitochondrial DNA Content, and Birth Weight in the INMA (Spain) and ENVIRONAGE (Belgium) Birth Cohorts