Quality of reporting in scientific publications: A report from the 2017 Peer Review Congress

Quality of reporting in scientific publications: A report from the 2017 Peer Review Congress

Hridey Manghwani, Manager, Corporate Communications, Editage, is attending the 2017 Peer Review Congress. This is Hridey’s private report of the 2nd day of the conference. 

After the keynote address by David Moher from Ottawa Hospital Research Institute, Day Two of the Eighth International Congress on Peer Review and Scientific Publishing switched gears with sessions on the quality of reporting.

Robert Frank from the University of Ottawa introduced his group’s probe that analyzed the variables associated with false results in imaging research. Meta analyses (MAs) are the closest proxy to the truth in imaging studies. Study-level (citation rate, publication order, and publication timing relative to STARD)and journal-level variables(Influence factor, STARD endorsement, and Cited half-life) were studied for their association with the distance inbetween primary results and summary estimates from MAs. After analyzing 98 MAs containing 1458 primary studies, the investigate concluded that the distance inbetween primary results and summary estimates from MAs was not smaller for studies published in high influence factor (IF) journals compared to lower IF journals. Their conclusions re-emphasized that higher IFs and citation rates are not indicative of better examine quality.

Next up was Sarah Daisy Kosa from McMaster University and Toronto General Hospital Health Network. Her investigate discussed discrepancies inbetween reported trial data and those updated in clinical trial registries for studies published in high-IF journals. Her research group searched PubMed for all randomized clinical trials (RCTs) published in the top Five (as per the 2014 IF published by Thomson Reuters) medicine journals inbetween 2012 and 2015. Of the 200 RCTs they studied, discrepancies were observed inbetween publications and clinical trial registries. Their investigate clearly highlighted the need for uniform reporting of clinical data in publications and registries in order to preserve the quality and integrity of scientific research.


Cole Wayant from Oklahoma State University Center for Health Sciences then took the stage to discuss the quality of methods and reporting of systematic reviews (SRs) underpinning clinical practice guidelines. His investigate analysed MAs and SRs based on PRISMA (Preferred Reporting Items for Systematic Reviews and Meta-analyses) and AMSTAR (A Measurement instrument to assess Systematic Reviews) guidelines. His examine noted that many items on these checklists were missing for MAs and SRs, reemphasizing the need for higher quality SRs, especially because they are considered level 1A evidence and can impair clinical decisions, thereby hindering evidence-based medicine.

The final and perhaps the most engaging presentation of the session was delivered by Kaveh Zakeri, a medical resident from the University of California, San Diego. His examine analyzed the existence of optimism bias in contemporary National Clinical Trials Network Phase Trio trials. Zakeri’s group identified 185 published studies from PubMed that were phase III randomized co-operative group clinical trials inbetween January 2007 and January 2017. By comparing the proposed and observed effect sizes, they calculated proposed-to-observed hazard ratios (HRs) that were indicative of therapy benefit and reduction in adverse effects. Their results helped them conclude that most NCI-sponsored Phase Trio trials could not demonstrate statistically significant benefits of fresh therapies. They suggested that albeit the magnitude of optimism bias had diminished compared to studies inbetween 1955 and 2006, better rationales for proposed effect sizes were required for clinical trial protocols. Many trials in very specialized fields cannot be repeated and even tho’ the sample size can be puny owing to an inability to recruit enough patients for a trial, most phase II and phase III trials provided no rationale at all for the proposed effect size.

We’ll cover the sessions of the 2nd half in the next part of the report. The post-lunch sessions will discuss scientific literature, trial registration, and funding/grant review. 



Related video: How Donald Trump Answers A Question

You may also like...

Leave a Reply

Your email address will not be published. Required fields are marked *