Indian Journal of Anaesthesia  
About us | Editorial board | Search | Ahead of print | Current Issue | Past Issues | Instructions
Home | Login  | Users Online: 1383  Print this pageEmail this pageSmall font sizeDefault font sizeIncrease font size    




 
 Table of Contents    
LETTERS TO EDITOR
Year : 2019  |  Volume : 63  |  Issue : 6  |  Page : 505-507  

Publication bias - Importance of studies with negative results!


Department of Anaesthesiology, Basavatarakam Indo-American Cancer Hospital and Research Institute, Hyderabad, Telangana, India

Date of Web Publication11-Jun-2019

Correspondence Address:
Dr. Abhijit S Nair
Department of Anaesthesiology, Basavatarakam Indo-American Cancer Hospital and Research Institute, Hyderabad - 500 034, Telangana
India
Login to access the Email id

Source of Support: None, Conflict of Interest: None


DOI: 10.4103/ija.IJA_142_19

Rights and Permissions

How to cite this article:
Nair AS. Publication bias - Importance of studies with negative results!. Indian J Anaesth 2019;63:505-7

How to cite this URL:
Nair AS. Publication bias - Importance of studies with negative results!. Indian J Anaesth [serial online] 2019 [cited 2019 Jun 27];63:505-7. Available from: http://www.ijaweb.org/text.asp?2019/63/6/505/259933



Sir,

Publication bias is defined as the failure to publish the results of a study on the basis of the direction or strength of the study findings.[1] This may mean that only studies which have statistically significant positive results get published and the statistically insignificant or negative studies does not get published. Of the several reasons of this bias the important ones are rejection (by editors, reviewers), lack of interest to revise, competing interests, lack of motivation to write in spite of conducting the study.[2] Many researchers do not publish research with negative results because they consider it as a failed research which is not true. If the hypothesis made by them is rejected based on results of a study with sound methodology, it does not mean it is a failed research.

There are three reasons for negative results: studies with small sample size and lacking power, no difference between groups, and more complications or adverse events in the study group. It is therefore obvious that either the editor does not send the research with negative results for further review or the reviewers reject the manuscript upfront.[3] An unpublished study with negative results also leads to a significant amount of monetary loss and time of the researchers and/or funding body involved. Clinical trials at various levels funded by pharmaceutical companies involving volunteers or patients which demonstrate adverse events does not get published in peer-reviewed journals for obvious reasons.

Systematic reviews and meta-analyses have an important place in modern day evidence-based clinical practice. Meta-analyses involve statistical analysis of pooled data of all the randomised controlled trials. However due to publication bias, the final analysis does not involve negative data as it has either not been published ever or has been rejected. Therefore, the practice guidelines that evolve from the results of systematic reviews and meta-analyses which comes under the category of level 1 evidence has to be taken into consideration with a pinch of salt.

Systematic reviews and meta-analyses use a funnel plot to check for the existence of publication bias or systematic heterogenicity in the studies taken for analysis. If the plot is symmetric inverted funnel shape, publication bias is unlikely.[4] If the funnel plot is asymmetric, it means that there is a systematic difference between studies of higher and lower precision [Figure 1]a and [Figure 1]b. Egger's regression is a statistical measure for quantifying funnel plot asymmetry.[5] Rosenthal's fail-safe number or “fail-safe N method” is another way of determining publication bias.[6] It identifies the number of additional negative studies to increase the P value in a meta-analysis to above 0.05. Although it a simple way of deriving a number, it is dependent on P value.
Figure 1: (a) Hypothetical symmetric funnel plot showing no publication bias. The figure also shows were to look for 95% confidence interval, overall effect, and study result. (b) An unlabeled hypothetical, asymmetric funnel plot showing publication bias

Click here to view


Preferred Reporting Items for Systematic Reviews and Meta-Analyses (PRISMA) statement and a flowchart of PRISMA is mandatory as it helps authors in improving reporting of systematic review and meta-analyses. Item no. 16 of the PRISMA checklist is titled 'Meta-bias(es)' where the authors need to specify if there was any planned assessment of meta-bias(es) like publication bias across studies or selective reporting within studies.

(http://www.prisma-statement.org/documents/PRISMA-P-checklist.pdf)

This item might not help in the analysis if there is a dearth of published or reported negative trials.

In conclusion, the editorial board should insist authors on submitting negative results also which should be considered for publication if found suitable based on appropriate methods, statistical methods used and acceptable discussion.

Financial support and sponsorship

Nil.

Conflicts of interest

There are no conflicts of interest.



 
   References Top

1.
DeVito NJ, Goldacre B. Catalogue of bias: Publication bias. BMJ Evid Based Med 2018. pii: bmjebm-2018-111107.  Back to cited text no. 1
    
2.
Mlinarić A, Horvat M, Šupak Smolčić V. Dealing with the positive publication bias: Why you should really publish your negative results. Biochem Med (Zagreb) 2017;27:030201.  Back to cited text no. 2
    
3.
Joober R, Schmitz N, Annable L, Boksa P. Publication bias: What are the challenges and can they be overcome? J Psychiatry Neurosci 2012;37:149-52.  Back to cited text no. 3
    
4.
Hedin RJ, Umberham BA, Detweiler BN, Kollmorgen L, Vassar M. Publication bias and nonreporting found in majority of systematic reviews and meta-analyses in anesthesiology journals. Anesth Analg 2016;123:1018-25.  Back to cited text no. 4
    
5.
Furuya-Kanamori, Barendregt JJ, Doi SAR. A new improved graphical and quantitative method for detecting bias in meta-analysis. Int J Evid Based Healthc 2018;16:195-203.  Back to cited text no. 5
    
6.
Fragkos KC, Tsagris M, Frangos CC. Publication bias in meta-analysis: Confidence intervals for Rosenthal's fail-safe number. Int Sch Res Notices 2014;2014:825383.  Back to cited text no. 6
    


    Figures

  [Figure 1]



 

Top
 
  Search
 
    Similar in PUBMED
   Search Pubmed for
   Search in Google Scholar for
    Access Statistics
    Email Alert *
    Add to My List *
* Registration required (free)  

 
  In this article
    References
    Article Figures

 Article Access Statistics
    Viewed195    
    Printed0    
    Emailed0    
    PDF Downloaded160    
    Comments [Add]    

Recommend this journal