Publication bias in animal research presented at the 2008 Society of Critical Care Medicine Conference
BMC Research Notes volume 10, Article number: 262 (2017)
To determine a direct measure of publication bias by determining subsequent full-paper publication (P) of studies reported in animal research abstracts presented at an international conference (A).
We selected 100 random (using a random-number generator) A from the 2008 Society of Critical Care Medicine Conference. Using a data collection form and study manual, we recorded methodology and result variables from A. We searched PubMed and EMBASE to June 2015, and DOAJ and Google Scholar to May 2017 to screen for subsequent P. Methodology and result variables were recorded from P to determine changes in reporting from A. Predictors of P were examined using Fisher’s Exact Test.
62% (95% CI 52–71%) of studies described in A were subsequently P after a median 19 [IQR 9–33.3] months from conference presentation. Reporting of studies in A was of low quality: randomized 27% (the method of randomization and allocation concealment not described), blinded 0%, sample-size calculation stated 0%, specifying the primary outcome 26%, numbers given with denominators 6%, and stating number of animals used 47%. Only being an orally presented (vs. poster presented) A (14/16 vs. 48/84, p = 0.025) predicted P. Reporting of studies in P was of poor quality: randomized 39% (the method of randomization and allocation concealment not described), likely blinded 6%, primary outcome specified 5%, sample size calculation stated 0%, numbers given with denominators 34%, and number of animals used stated 56%. Changes in reporting from A to P occurred: from non-randomized to randomized 19%, from non-blinded to blinded 6%, from negative to positive outcomes 8%, from having to not having a stated primary outcome 16%, and from non-statistically to statistically significant findings 37%. Post-hoc, using publication data, P was predicted by having positive outcomes (published 62/62, unpublished 33/38; p = 0.003), or statistically significant results (published 58/62, unpublished 20/38; p < 0.001).
Only 62% (95% CI 52–71%) of animal research A are subsequently P; this was predicted by oral presentation of the A, finally having positive outcomes, and finally having statistically significant results. Publication bias is prevalent in critical care animal research.
Publication bias (PB) refers to preferential publication of research findings that have statistically significant positive outcomes . This is problematic because the published literature is what scientists and clinicians use to inform their research and clinical practice. Indeed, the enterprise of evidence based medicine rests on the assumption that the published literature is an accurate representation of the current knowledge base. If evidence from systematic literature reviews is a biased representation of knowledge, this is a heavy blow to the paradigm of evidence based medicine . The problem of PB is well recognized in clinical research; reviews suggest that only about half of research is published, and positive findings are a strong predictor of publication [3, 4].
Biomedical animal research (AR) is used to inform human research and practice on the assumption that animals are causal analogical models of human physiology and response to drugs and disease . If PB is prevalent in AR, this is problematic for several reasons. First, evidence from published AR would be a biased representation of pre-clinical knowledge and thus be potentially misleading and dangerous for informing human medicine [6, 7]. Second, the moral permissibility of AR, based on the claim that the harm to sentient animals used in AR is outweighed by large human benefit from this AR would be unacceptably weakened . For example, animals harmed in unpublished AR cannot contribute to biomedical knowledge, and thus were harmed for no good reason. Moreover, the animals used in unpublished AR that had negative findings were harmed with additional potential harm also to humans, due to subsequent biomedical research and practice following false leads or safety data from the published AR . Some data indirectly suggests PB in AR is common. Using statistical tools that examine small study bias (a larger effect size in small studies compared to larger studies), several AR reviews suggest PB occurs [10,11,12,13,14,15,16,17]. Another recent study found AR suffers from excess significance bias, again indirectly suggesting PB . Of 174 basic science study abstracts submitted to a major gastroenterology conference, of which 90 were “animal studies”, subsequent publication occurred for 47%, with no statistically significant predictors of publication found; however, results for AR were not reported separately . To our knowledge there has not been a direct assessment of whether PB occurs in AR.
Here we aimed to determine whether AR abstracts presented at an international critical care conference are subsequently published in the peer-reviewed literature, and what the predictors of subsequent publication may be. We chose critical care because inducing critical illness may be one of the most invasive fields of AR. In addition, our method allows a direct assessment of PB of AR that has reached the stage of warranting presentation to peers.
The University of Alberta Health Research Ethics Board waived the requirement for review or consent because the study involved only publicly available data, and no individual that would require consent to participation.
We reviewed 100 random published abstracts of AR from the 2008 Society of Critical Care Medicine international conference . The abstracts were chosen using a random number generator. There were no restrictions other than that the abstract reported an AR experiment, defined as: a procedure for collecting scientific data on the response to an intervention in a systematic way to maximize the chance of answering a question correctly or to provide material for the generation of new hypotheses . If there was any doubt about inclusion, this was discussed among the two authors to achieve consensus. A data collection form and instruction manual (see Additional files 1, 2) were created based on published Canadian (the Canadian Council on Animal Care in science guidelines “Choosing an appropriate endpoint in experiments using animals for research, teaching and testing” and “Animal use protocol review”), United States (the Institute for Laboratory Animal Research, National Research Council “Guidance for the description of animal research in scientific publications”), and United Kingdom (the National Center for the 3Rs “Animal Research: Reporting of In Vivo Experiments” ARRIVE guidelines) recommendations for reporting AR [22,23,24,25]. These guidelines were used as they are comprehensive, well referenced, readily available, and based upon literature review. For example, the ARRIVE guidelines were developed to improve the quality of reporting AR, and are endorsed by over 100 journals from all over the world . Data were obtained for factors important to methodological quality.
The data collection form was completed for all 100 critical care AR (using mammals) abstracts. Both authors independently completed forms for the first 10 abstracts, discussing the data after every fifth abstract until consistent agreement was obtained. Thereafter, one author completed forms on all abstracts, and the other author independently did so for every fifth abstract (with discussion of the data to maintain consistent agreement), and for any data considered uncertain (with discussion until consensus). The instruction manual made clear definitions for all data collection; for example, a sample size calculation was defined as describing, for the primary outcome, a p value (alpha), power (1-beta), and minimally important difference (the difference between groups that the study is powered to detect).
Search for publication
After the abstract data was finalized in the data collection form, we searched both PubMed and EMBASE to determine subsequent publication of the data presented in the abstracts. The search strategy was defined in the instruction manual. We searched for the first, second, or last author; and at least two MeSH subject terms (e.g., acute lung injury, or sepsis). The search strategy was developed to be as inclusive as possible, by using the operator “OR” between the authors searched and between the MeSH terms used. Abstracts that were later published as only a small part of a larger study were considered published. All titles and abstracts of identified publications were reviewed to determine if they reported the study from the abstract. To be considered publication of the abstract results the published paper needed to report the same experiment, and we included the publication even if more animals were used and/or more experiments were done than reported in the abstract. If there was any doubt, both authors together discussed the titles and abstracts and reached consensus. If publication could not be found by one of the authors, the second author also independently searched to confirm this; if publication was then found, both authors discussed the finding and reached consensus. The search was done from 2007 (up to 14 months prior to the conference date) to June 2015 (up to 86 months after the conference date) inclusive.
In response to an anonymous reviewer, in May 2017 one author (ARJ) searched DOAJ (the Directory of Open Access Journals) and Google Scholar for subsequent publication of any of the abstracts determined unpublished by the above strategy. The search strategy was modified to be consistent with the PubMed and EMBASE searches. On DOAJ we searched separately for each of the first or last author, and if more than 30 publications were listed, we added (separately, one at a time) each of two subject terms from the abstract title. On Google Scholar we searched for each of the first or last author, added (separately, one at a time) each of two subject terms from the abstract title into the “contains all of the words” box, and limited the search to from years 2007 to 2017. We screened all returned titles, and if necessary, abstracts and full text, up to and including 5 pages of each search result.
Finally, as suggested by another reviewer, we emailed an author of each unpublished abstract to ask if the abstract had been published in full, and if so, to provide the citation. For each abstract, one author (ARJ) searched PubMed for the abstract author (starting with first author, and if no email found, last author, second author, second last author, etc.) in order to obtain a corresponding author email address. To be sure this was a current email for the correct abstract author, the address was to be from a publication since 2010, and on a topic related to the abstract, preferably with a same coauthor(s) as in the abstract. If the email was returned undeliverable, the search was repeated to obtain another of the abstract authors’ email. For the 3 emails that remained undeliverable we identified a current corresponding co-author of a publication with the target author, and asked this contact to forward the request.
A data collection form and instruction manual (see Additional files 1, 2) were used for each identified publication. Both authors independently completed data collection forms for the first 5 identified publications, discussing the data until consistent agreement was obtained. Thereafter, one author completed forms on all publications, and the other author independently did so for every fifth publication (with discussion of the data to maintain consistent agreement), and for any data considered uncertain (with discussion until consensus). The instruction manual made clear definitions for all data collection variables.
Data are presented using descriptive statistics, and were analyzed using SPSS. The primary outcome was pre-specified as subsequent publication of AR abstracts, with 95% adjusted Wald Confidence Intervals (CI). Assuming the subsequent publication rate of AR will be similar to that for clinical research weighted mean full publication rate of 44.5% (95% CI 43.9–45.1%) , an abstract sample of n = 80 will provide reasonable 95% CI of ±10%. Thus, we decided to include 80 abstracts, and if the timing allows, add a random sample of 20 abstracts to increase the sample size prior to any data analysis. We compared abstracts that were published to those that were not using Chi square and Fisher’s Exact test, with statistical significance accepted at p < 0.05 without correction for multiple comparisons. Pre-defined possible predictors included methodology (randomization, blinding, sample size calculation, primary outcome specification, results reported with denominators, number of animals used stated), ethical (number of animals used >19, highest species used rodent vs other), outcome (positive finding, statistically significant finding), and type of animal model (sepsis, drug used, surgery performed, animals stated to be killed) variables. Post-hoc we tested for possible predictors including methodology (randomization, blinding) and outcome (positive finding, statistically significant finding) variables updated with information from publication if available. We also determined whether there were changes in important information reported between the abstract and subsequent publication, using our a priori definitions in the data collection form and instruction manual.
61 [61% (95% CI 51–70%)] abstracts were subsequently published after a median 19 [IQR 9–33, range 0–68] months by searching PubMed and EMBASE. The search of DOAJ found no publications, and Google Scholar found one further publication, for a publication rate of 62% (95% CI 52–71%). Publication was usually within 3 years; months to publication was mean 22.6 (SD 17.1), median 19 [IQR 9, 33.3], range 0–68 months; the 80th percentile was 35 months (2.9 years); 90th percentile 52 months (4.3 years); and 95th percentile 60 months (5 years).
Emails were sent on May 25 (and to non-responders again on June 1) to an abstract author for the initially determined 39 unpublished abstracts. There were 13 replies by June 8 (within 2 weeks): 9 confirmed non-publication, and 4 claimed publication. However, the citations provided in the 4 clearly did not match the abstract methodology despite being on a similar topic (1 was published 4 years before the abstract and used different study interventions; 1 was a global head injury model whereas the abstract was of a focal brain injury model; 1 randomized to different intervention groups than in the abstract; and 1 was published 12 years before the abstract and used a different animal model). Thus, no publication was identified by this emailing strategy.
Predictors of abstract publication
The differences between published and unpublished abstracts are given in Table 1. Methodological quality of reporting in the abstracts was poor, with randomization in 27 (27%; none reported the method of randomization or allocation concealment), blinding in 0 (0%), sample size calculation in 0 (0%), primary outcome stated in 26 (26%), number with denominators given for main outcomes in 6 (6%), and number of animals used stated in 47 (47%; median 18, IQR 11–24, range 1–60; total 957). Most abstracts reported mainly positive outcomes (90, 90%), and these outcomes were statistically significant in 55 (55%). The only statistically significant predictor of subsequent publication was being an oral presentation at the conference (p = 0.024).
Changes from abstract to publication
Changes in reporting between the abstract and publication are given in Table 2. The reported methodological quality of publications was poor: randomized in 24 (39%; method of randomization and allocation concealment not reported), blinding in 4 (6%; it was unclear for which outcomes blinding occurred in all of these), sample size calculation in 0 (0%), primary outcome stated in 3 (5%), numbers given with denominators in results 21 (34%), and number of animals stated in 35 (56%; median 20, IQR 14–35, range 5–125, total 993). Changes in reporting between abstract and publication included: from non-randomized in the abstract to randomized in the publication for 12 (19%), from no mention of blinding to blinding for some outcomes in 4 (6%), from having a primary outcome in the abstract to no primary outcome stated in the publication for 10 (16%; this is for 10/13, 77% of the abstracts that stated a primary outcome), from main outcomes being negative in the abstract to being positive (or excluded) in the publication for 5 (8%), and from main outcomes being non-statistically significant (or no statistical significance stated) in the abstract to being statistically significant in the publication for 23 (37%). In the publications the main outcomes reported from the abstract were positive for 62 (100%), and statistically significantly so for 58 (94%).
Post-hoc predictors of subsequent publication
Given the unexpected changes from abstract to publication, we determined predictors for subsequent publication using updated numbers combining abstract and publication (when available) data (Table 1). The methodological variables of randomization or blinding did not predict publication, but finally having main outcomes being positive (p = 0.003), and finally having statistically significant main outcomes (p < 0.001) were predictors of publication.
Post-hoc comparison of oral versus poster presentations
Given that oral presentation was a possible predictor of subsequent publication (p = 0.025), we compared oral to poster presentations on quality of abstracts and publications (Table 3). Orally presented abstracts were less likely to report animal numbers (p = 0.02) and randomization (p = 0.06) than poster presented abstracts, and orally presented abstracts that were subsequently published remained less likely to report animal numbers and randomization than poster presentations that were subsequently published.
Post-hoc comparison of abstracts and publications according to journal impact factor
Given the lack of methodological differences between published and unpublished abstracts, as suggested by a reviewer, we compared quality between lower (at or below median) and higher (above median) journal impact factor publications. There were no statistically significant differences in abstract quality between those subsequently published in lower vs higher impact journals (Table 4). The only statistically significant difference in quality between publications in lower vs higher impact journals was in blinding for some outcomes, which was used in 0% of lower impact publications and 4 (13%) of higher impact publications (p = 0.049).
There are several important findings from this study. First, only 62% (95% CI 52–71%) of AR abstracts presented at an international critical care conference, the Society for Critical Care Medicine 2008 Conference, were subsequently published. This means that much AR that is ready for presentation to peers does not contribute to the biomedical literature. Second, predictors of publication were oral presentation, and finally having positive outcomes and statistically significant results. Orally presented abstracts were not of higher methodological or ethical quality. These predictors confirm that much of the reason for non-publication is due to PB, which leads to a biased representation of biomedical research in the published literature. This is the first study of which we are aware to directly confirm PB in the AR literature. Third, the reported methodological quality of abstracts and publications is poor, particularly due to no mention of method of randomization, allocation concealment, or sample size calculation, and infrequent mention of randomization, primary outcomes, number of animals used, and numbers with denominators provided for results. This suggests that even published AR has poor internal validity. That the methodological quality did not differ between abstracts and publications suggests that this quality was not a determinant of acceptance for publication. Moreover, quality did not differ between abstracts subsequently published in higher versus lower impact journals. Finally, the change in reporting from abstract to publication is of concern. Although infrequent, the increase in reporting of randomization (by 19%) and blinding (by 6%), the decrease in reporting of a primary outcome (by 16%), and the increase in positive findings (by 8%) and statistically significant findings (by 37%) suggest that in published AR there may be selective analysis and outcome reporting bias that aims to find reportable results. The change in numbers of animals reported from abstract to publication in 13/35 (37%) of those studies reporting animal numbers [in the publication the number was smaller in 3 (9%) and larger in 10 (29%)] also suggests the possibility of checking the data during the ongoing study until a significant difference is found (i.e., repeated interim analyses of data as they accumulate). These changes were not significantly less frequent in publications in higher vs. lower impact factor journals (Table 5).
These findings are compatible with previous research. In clinical research, PB in reporting findings from abstracts presented at conferences has been well recognized; only about 45% of abstracts are subsequently published, and predictors of subsequent publication include positive and statistically significant findings [2,3,4]. In addition, for clinical research, the quality of poster and orally presented abstracts has been similar . Previous indirect assessments of PB in AR have suggested small study bias, and excess significance bias [10,11,12,13,14,15,16,17,18]. A survey of animal researchers in the Netherlands found that respondents estimated only 50% (95% CI 32–70%) “of ethics-approved experiments performed in experimental animal research is published,” well within our 95% CI of publication rate . Many previous reviews have found poor methodological quality of reported AR, including in the field of critical care [28,29,30,31,32,33]. Although selective analysis and outcome reporting bias, and data “peaking” to find significant results (instead of pre-defined sample sizes being calculated) have been suspected in AR by commentators, this is the first confirmation of this occurring as far as we are aware [6, 7, 18, 34].
These findings are important for several reasons. In clinical research PB poses a serious threat to validly assessing the effectiveness of new therapies in systematic reviews; research subjects have participated in a study without (or even negatively) contributing to science; and those who make health care decisions are faced with a biased subset of scientific evidence [2,3,4]. PB in pre-clinical AR is important for the same reasons; indeed, AR is often used to guide decisions for further AR and for human clinical research, and the ethical justification for AR is exactly this contribution to making decisions of benefit for human clinical research and treatment [2, 6, 7, 9]. The finding of PB in AR suggests that the valid extrapolation of findings from the AR literature to inform research and treatment in humans is seriously threatened. There are several important implications. First, strategies to combat PB are a priority. For example, funding agencies that support AR could tie funding to subsequent full publication of results, and academic institutions that employ scientists doing AR could tie full publication of results to academic review . In addition, journal editors and referees have a responsibility to publish negative results, if of good quality. Second, strategies to identify PB are a priority, for example, international registration of AR protocols [2, 6, 7]. Third, attention to methodological quality requirements in funding, ethics review, and academic review decisions is a priority [28,29,30,31,32,33]. For example, if sample size calculations and statistical design for primary outcomes are required in protocols and grant applications, this can prevent data dredging, selective outcome reporting bias, ongoing recruitment aiming to obtain statistically significant outcomes, and other biases .
There are limitations to this study. First, it is possible that some abstracts were published but this was not detected by our search strategy. Second, it is also possible that methodological quality was better than reported, particularly in abstracts where space limitations are restrictive. We believe that this explanation is problematic. Optimal methods of randomization, allocation concealment, and blinding (of the experiments and assessment of their outcomes) are time consuming and expensive to implement. Randomization, blinding, sample size calculation for a pre-specified primary outcome (limiting “sampling to a foregone conclusion” and multiple statistical testing for post-hoc outcomes), and describing research subject numbers (including attrition or exclusion of animals) markedly improve the quality, validity, and reproducibility of even preliminary experimental results. Not reporting this information can make published findings unreliable, regardless of whether the information was in fact known to the authors. This is why guidelines for reporting clinical research in conference abstracts include these points in their minimal standards [36,37,38]. This is also in the spirit of the ARRIVE guidelines, meant to ensure publications are “’fit for purpose,’ meaning that incomplete reporting of relevant information effectively renders many publications of limited value as instruments to inform policy or clinical and scientific practice… unusable due to poor reporting.”  These points are also in the “core set of research parameters” in the US National Institute of Neurological Disorders and Stroke workshop statement, those “that should be addressed when reporting the results of animal experiments… that would allow informed judgment about the findings” . Third, the sample size of 100 abstracts and 62 publications limits study power to detect predictors of publication, or of publication in lower versus higher impact journals. The analyses of predictors should be considered exploratory only, particularly given the multiple statistical testing. Fourth, this study is limited to abstracts reported at one international conference, and in the field of critical care, and may not generalize to other conferences and fields. Finally, we did not assess the quality of statistical methods used; however, this may not be possible given that most animal research publications examine statistically well over 20 outcomes .
Nevertheless, the strengths of this study mitigate some of these concerns. This study was done with clear data definitions and a data collection form, and with a clear rigorous search strategy used by two authors independently. In addition, the field of critical care AR usually involves quite invasive management of animals (e.g., invasive procedures done for animal monitoring, and invasive study interventions such as surgery and ventilation), and therefore might be expected to be most likely to be reported and done to high standards . Finally, the results are similar to many previous reports suggesting PB in clinical and AR, and reports regarding the methodological quality of AR [2,3,4, 6, 7, 10,11,12,13,14,15,16,17,18, 27,28,29,30,31,32,33]. Our method of determining PB, that is, by determining subsequent publication of data presented in abstract form at an international conference, has been widely accepted in clinical research [3, 4].
A direct assessment of subsequent publication of AR presented as abstracts at an international critical care conference confirms PB is prevalent. This is of concern because it suggests that published AR is a biased representation of research findings, and that animals are harmed without benefit, and with potential harm (misleading literature reviews), to humans. If AR has any chance of translating to human medicine addressing this problem of PB must be a high priority for researchers, funders, journals, and clinicians alike [39, 40].
Guyatt GH, Oxman AD, Montori V, Vist G, Kunz R, Brozek J, et al. GRADE guidelines: 5. Rating the quality of evidence-publication bias. J Clin Epidemiol. 2011;64(12):1277–82.
Chan A, Song F, Vickers A, Jefferson T, Dickersin K, Gotzsche PC, et al. Increasing value and reducing waste: addressing inaccessible research. Lancet. 2014;383(9913):257–66.
Scherer RW, Langenberg P, von Elm E. Full publication of results initially presented in abstracts. Cochrane Database Syst Rev. 2007;2:MR000005. doi:10.1002/14651858.MR000005.pub3.
Song F, Parekh-Bhurke S, Hooper L, Loke YK, Ryder JJ, Sutton AJ, Hing CB, Harvey I. Extent of publication bias in different categories of research cohorts: a meta-analysis of empirical studies. BMC Med Res Methodol. 2009;9:79.
Greek R, Rice MJ. Animal models and conserved processes. Theor Biol Med Model. 2012;9:40.
van der Worp HB, Howells DW, Sena ES, Porritt MJ, Rewell S, O’Collins V, Macleod MR. Can animal models of disease reliably inform human studies? PLoS Med. 2010;7(3):e1000245.
Green SB. Can animal data translate to innovations necessary for a new era of patient-centred and individualized healthcare? Bias in preclinical animal research. BMC Med Ethics. 2015;15:53.
Garrett JR. The ethics of animal research: an overview of the debate. In: Garrett JR, editor. The ethics of animal research: exploring the controversy. MIT: USA; 2012. p. 1–16.
Bass R. Lives in the balance: utilitarianism and animal research. In: Garrett JR, editor. The ethics of animal research: exploring the controversy. MIT: USA; 2012. p. 81–105.
Antonic A, Sena ES, Lees JS, Wills TE, Skeers P, Batchelor PE, Macleod MR, Howells DW. Stem cell transplantation in traumatic spinal cord injury: a systematic review and meta-analysis of animal studies. PLoS Biol. 2013;11(12):e1001738.
Watzlawick R, Sena ES, Dirnagl U. Effect of reporting bias of RhoA/ROCK-Blockade intervention on locomotor recovery after spinal cord injury: a systematic review and meta-analysis. JAMA Neurol. 2014;71(1):91–9.
Sena ES, van der Worp HB, Bath PM, Howells DW, Macleod MR. Publication bias in reports of animal stroke studies leads to major overstatement of efficacy. PLoS Biol. 2010;8:e1000344.
Pedder H, Vesterinen HM, Macleod MR, Wardlaw JM. Systematic review and meta-analysis of interventions tested in animal models of lacunar stroke. Stroke. 2014;45:563–70.
Macleod MR, O’Collins T, Howells DW, Donnan GA. Pooling of animal experimental data reveals influence of study design and publication bias. Stroke. 2004;35:1203–8.
Sena ES, Briscoe CL, Howells DW, Donnan GA, Sandercook PA, Macleod MR. Factors affecting the apparent efficacy and safety of tissue plasminogen activator in thrombotic occlusion models of stroke: systematic review and meta-analysis. J Cereb Blood Flow Metab. 2010;30:1905–13.
Kwon BK, Okon EB, Tsai E, Beattie MS, Bresnahan JC, Magnuson DK, et al. A grading system to evaluate objectively the strength of pre-clinical data of acute neuroprotective therapies for clinical translation in spinal cord injury. J Neurotrauma. 2011;28(8):1525–43.
Anglemyer AT, Krauth D, Bero L. Industry sponsorship and publication bias among animal studies evaluating the effects of statins on atherosclerosis and bone outcomes: a meta-analysis. BMC Med Res Methodol. 2015;15:12.
Tsilidis KK, Panagiotou OA, Sena ES, Aretouli E, Evangenlou E, Howells DW, et al. Evaluation of excess significance bias in animal studies of neurological diseases. PLoS Biol. 2013;11(7):e1001609.
Timmer A, Hilsden RJ, Cole J, Hailey D, Sutherland LR. Publication bias in gastroenterological research—a retrospective cohort study based on abstracts submitted to a scientific meeting. BMC Med Res Methodol. 2002;2:7.
Society of Critical Care Medicine’s 37th Critical Care Congress Abstracts. Crit Care Med. 2007;35(12 Suppl):A1–A289.
Festing MF, Altman DG. Guidelines for the design and statistical analysis of experiments using laboratory animals. ILAR J. 2002;43(4):244–58.
Kilkenny C, Browne WJ, Cuthill IC, Emerson M, Altman DG. Improving bioscience research reporting: the ARRIVE guidelines for reporting animal research. PLoS Biol. 2010;8(6):e1000412.
Institute for Laboratory Animal Research, National Research Council. Guidance for the description of animal research in scientific publications. Washington DC: National Academy of Sciences; 2011.
Canadian Council on Animal Care in science. CCAC Guidelines on: choosing an appropriate endpoint in experiments using animals for research, teaching and testing. Ottawa, ON, Canada: Canadian Council on Animal Care; 1998. http://www.ccac.ca/Documents/Standards/Guidelines/Appropriate_endpoint.pdf. Accessed 10 Sep 2016.
Canadian Council on Animal Care in science: CCAC Guidelines on: animal use protocol review. Ottawa, ON, Canada: Canadian Council on Animal Care; 1997. http://www.ccac.ca/Documents/Standards/Guidelines/Protocol_Review.pdf. Accessed 10 Sep 2016.
Knobloch K, Yoon U, Rennekampff HO, Vogt PM. Quality of reporting according to the CONSORT, STROBE and Timmer instrument at the American Burn Association (ABA) annual meetings 2000 and 2008. BMC Med Res Methodol. 2011;11:161.
ter Riet G, Korevaar DA, Leenaars M, Sterk PJ, Van Noorden CFJ, Bouter LM, et al. Publication bias in laboratory animal research: a survey on magnitude, drivers, consequences, and potential solutions. PLoS ONE. 2012;7:e43404.
Bara M, Joffe AR. The methodological quality of animal research in critical care: the public face of science. Ann Intensiv Care. 2014;4:26.
Killkenny C, Parsons N, Kadyszewski E, Festing MFW, Cuthill IC, Fry D, Hutton J, Altman DG. Survey of the quality of experimental design, statistical analysis and reporting of research using animals. PLoS ONE. 2009;4(11):e7284.
Sena E, van der Worp B, Howells D, Macleod M. How can we improve the preclinical development of drugs for stroke? Trends Neurosci. 2007;30:433–9.
Baker D, Lidster K, Sottomayor A, Amor S. Two years later: journals are not yet enforcing the ARRIVE guidelines on reporting standards for pre-clinical animal studies. PLoS Biol. 2014;12(1):e1001756.
Landis SC, Amara SG, Asadullah K, Austin CP, Blumenstein R, Bradley EW, et al. A call for transparent reporting to optimize the predictive value of preclinical research. Nature. 2012;490:187–91.
Hawkes N. Poor quality animal studies cause clinical trials to follow false leads. BMJ. 2015;351:h5453.
Ludbrook J. Interim analyses of data as they accumulate in laboratory experimentation. BMC Med Res Methodol. 2003;3:15.
Hooijmans CR, Rovers MM, de Vries RBM, Leenaars M, Ritskes-Hoitinga M, Langendam MW. SYRCLE’s risk of bias tool for animal studies. BMC Med Res Methodol. 2014;14:43.
Hopewell S, Clarke M, Moher D, Wager E, Middleton P, Altman DG, Schulz KF, CONSORT Group. CONSORT for reporting randomized trials in journal and conference abstracts: explanation and elaboration. PLoS Med. 2008;5(1):e20.
Beller EM, Glasziou PP, Altman DG, Hopewell S, Bastian H, Chalmers I, Gotzsche PC, Lasserson T, Tovey D, for the PRISMA for Abstracts Group. PRISMA for Abstracts: reporting systematic reviews in journal and conference abstracts. PLOS Med. 2013;10(4):e1001419.
First draft of the STROBE checklist of items to be included when reporting observational studies in conference abstracts. 2017. https://www.strobe-statement.org/fileadmin/Strobe/uploads/checklists/STROBE_checklist_conference_abstract_DRAFT.pdf. Accessed 1 June 2017.
Pound P, Bracken MB. Is animal research sufficiently evidence based to be a cornerstone of biomedical research? BMJ. 2014;348:g3387.
Horrobin DF. Modern biomedical research: an internally self-consistent universe with little contact with medical reality. Nat Rev Drug Discov. 2003;2:151–4.
ARJ contributed to conception and design, acquisition, analysis and interpretation of data, and drafted the paper, and read and had final approval of the version to be published. UC contributed to design, acquisition and interpretation of data, and revising the manuscript critically for intellectual content, and read and had final approval of the version to be published. ARJ had full access to all the data in the study and takes responsibility for the integrity of the data and the accuracy of the data analysis. Both authors read and approved the final manuscript.
The authors declare that they have no competing interests.
Availability of data and materials
The dataset analyzed during the current study are available from the corresponding author on reasonable request.
Ethics approval and consent to participate
The University of Alberta Health Research Ethics Board waived the requirement for review or consent because the study involved only publicly available data, and no individual that would require consent to participation.
UC was supported for this research by a University of Alberta Office of the Provost and VP (Academic) Summer Research Award; the funding agency had no role in the design and conduct of the study; collection, management, analysis or interpretation of the data; preparation, review, or approval of the manuscript; or the decision to submit the manuscript for publication.
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Additional file 2. The study manual. This file provides the definitions used, and the details of methods for performing this study.
About this article
Cite this article
Conradi, U., Joffe, A.R. Publication bias in animal research presented at the 2008 Society of Critical Care Medicine Conference. BMC Res Notes 10, 262 (2017). https://doi.org/10.1186/s13104-017-2574-0
- Animal experimentation
- Bias (epidemiology)
- Intensive care
- Publication bias