Essay Essays are opinion pieces on a topic of broad interest to a general medical audience.

See all article types »

Why Current Publication Practices May Distort Science

  • Neal S Young mail,

    To whom correspondence should be addressed. E-mail:

  • John P. A Ioannidis,
  • Omar Al-Ubaydli
  • Published: October 07, 2008
  • DOI: 10.1371/journal.pmed.0050201

This essay makes the underlying assumption that scientific information is an economic commodity, and that scientific journals are a medium for its dissemination and exchange. While this exchange system differs from a conventional market in many senses, including the nature of payments, it shares the goal of transferring the commodity (knowledge) from its producers (scientists) to its consumers (other scientists, administrators, physicians, patients, and funding agencies). The function of this system has major consequences. Idealists may be offended that research be compared to widgets, but realists will acknowledge that journals generate revenue; publications are critical in drug development and marketing and to attract venture capital; and publishing defines successful scientific careers. Economic modelling of science may yield important insights (Table 1).

The Winner's Curse

In auction theory, under certain conditions, the bidder who wins tends to have overpaid. Consider oil firms bidding for drilling rights; companies estimate the size of the reserves, and estimates differ across firms. The average of all the firms' estimates would usually approximate the true reserve size. Since the firm with the highest estimate bids the most, the auction winner systematically overestimates, sometimes so substantially as to lose money in net terms [1]. When bidders are cognizant of the statistical processes of estimates and bids, they correct for the winner's curse by shading their bids down. This is why experienced bidders sometimes avoid the curse, as opposed to inexperienced ones [1–4]. Yet in numerous studies, bidder behaviour appears consistent with the winner's curse [5–8]. Indeed, the winner's curse was first proposed by oil operations researchers after they had recognised aberrant results in their own market.


The current system of publication in biomedical research provides a distorted view of the reality of scientific data that are generated in the laboratory and clinic. This system can be studied by applying principles from the field of economics. The “winner's curse,” a more general statement of publication bias, suggests that the small proportion of results chosen for publication are unrepresentative of scientists' repeated samplings of the real world. The self-correcting mechanism in science is retarded by the extreme imbalance between the abundance of supply (the output of basic science laboratories and clinical investigations) and the increasingly limited venues for publication (journals with sufficiently high impact). This system would be expected intrinsically to lead to the misallocation of resources. The scarcity of available outlets is artificial, based on the costs of printing in an electronic age and a belief that selectivity is equivalent to quality. Science is subject to great uncertainty: we cannot be confident now which efforts will ultimately yield worthwhile achievements. However, the current system abdicates to a small number of intermediates an authoritative prescience to anticipate a highly unpredictable future. In considering society's expectations and our own goals as scientists, we believe that there is a moral imperative to reconsider how scientific data are judged and disseminated.

An analogy can be applied to scientific publications. As with individual bidders in an auction, the average result from multiple studies yields a reasonable estimate of a “true” relationship. However, the more extreme, spectacular results (the largest treatment effects, the strongest associations, or the most unusually novel and exciting biological stories) may be preferentially published. Journals serve as intermediaries and may suffer minimal immediate consequences for errors of over- or mis-estimation, but it is the consumers of these laboratory and clinical results (other expert scientists; trainees choosing fields of endeavour; physicians and their patients; funding agencies; the media) who are “cursed” if these results are severely exaggerated—overvalued and unrepresentative of the true outcomes of many similar experiments. For example, initial clinical studies are often unrepresentative and misleading. An empirical evaluation of the 49 most-cited papers on the effectiveness of medical interventions, published in highly visible journals in 1990–2004, showed that a quarter of the randomised trials and five of six non-randomised studies had already been contradicted or found to have been exaggerated by 2005 [9]. The delay between the reporting of an initial positive study and subsequent publication of concurrently performed but negative results is measured in years [10,11]. An important role of systematic reviews may be to correct the inflated effects present in the initial studies published in famous journals [12], but this process may be similarly prolonged and even systematic reviews may perpetuate inflated results [13,14].


Table 1. Economic Terms and Analogies in Scientific Publication


More alarming is the general paucity in the literature of negative data. In some fields, almost all published studies show formally significant results so that statistical significance no longer appears discriminating [15,16]. Discovering selective reporting is not easy, but the implications are dire, as in the “hidden” results for antidepressant trials [17]: in a recent paper, it was shown that while almost all trials with “positive” results on antidepressants had been published, trials with “negative” results submitted to the US Food and Drug Administration, with few exceptions, remained either unpublished or were published with the results presented so that they would appear “positive” [17]. Negative or contradictory data may be discussed at conferences or among colleagues, but surface more publicly only when dominant paradigms are replaced. Sometimes, negative data do appear in refutation of prominent claims. In the “Proteus phenomenon”, an extreme result reported in the first published study is followed by an extreme opposite result; this sequence may cast doubt on the significance, meaning, or validity of any of the results [18]. Several factors may predict irreproducibility (small effects, small studies, “hot” fields, strong interests, large databases, flexible statistics) [19], but claiming that a specific study is wrong is a difficult, charged decision.

In the basic biological sciences, statistical considerations are secondary or nonexistent, results entirely unpredicted by hypotheses are celebrated, and there are few formal rules for reproducibility [20,21]. A signalling benefit from the market—good scientists being identified by their positive results—may be more powerful in the basic biological sciences than in clinical research, where the consequences of incorrect assessment of positive results are more dire. As with clinical research, prominent claims sometimes disappear over time [21]. If a posteriori considerations are met sceptically in clinical research, in basic biology they dominate. Negative data are not necessarily different than positive results as related to considerations of experimental design, execution, or importance. Much data are never formally refuted in print, but most promising preclinical work eventually fails to translate to clinical benefit [22]. Worse, in the course of ongoing experimentation, apparently negative studies are abandoned prematurely as wasteful.


Successful publication may be more difficult at present than in the past. The supply and demand of scientific production have changed. Across the health and life sciences, the number of published articles in Scopus-indexed journals rose from 590,807 in 1997 to 883,853 in 2007, a modest 50% increase. In the same decade, data acquisition has accelerated by many orders of magnitude: as an example, the current Cancer Genome Atlas project requires 10,000 times more sequencing effort than the Human Genome Project, but is expected to take a tenth of the time to complete [23]. In the current environment, the distinction between raw data and articles (telling for sure what more an article has compared with raw data) can sometimes become difficult. Only a small proportion of the explosively expanded output of biological laboratories appears in the modestly increased number of journal slots available for its publication, even if more data can be compacted in the average paper now than in the past.

Constriction on the demand side is further exaggerated by the disproportionate prominence of a very few journals. Moreover, these journals strive to attract specific papers, such as influential trials that generate publicity and profitable reprint sales. This “winner-take-all” reward structure [24] leaves very little space for “successful publication” for the vast majority of scientific work and further exaggerates the winner's curse. The acceptance rate decreases by 5.3% with doubling of circulation, and circulation rates differ by over 1,000-fold among 114 journals publishing clinical research [25]. For most published papers, “publication” often just signifies “final registration into oblivion”. Besides print circulation, in theory online journals should be readily visible, especially if open access. However, perhaps unjustifiably, most articles published in online journals remain rarely accessed. Only 73 of the many thousands of articles ever published by the 187 BMC-affiliated journals had over 10,000 accesses through their journal Web sites in the last year [26].

Impact factors are widely adopted as criteria for success, despite whatever qualms have been expressed [27–32]. They powerfully discriminate against submission to most journals, restricting acceptable outlets for publication. “Gaming” of impact factors is explicit. Editors make estimates of likely citations for submitted articles to gauge their interest in publication. The citation game [33,34] has created distinct hierarchical relationships among journals in different fields. In scientific fields with many citations, very few leading journals concentrate the top-cited work [35]: in each of the seven large fields to which the life sciences are divided by ISI Essential Indicators (each including several hundreds of journals), six journals account for 68%–94% of the 100 most-cited articles in the last decade (Clinical Medicine 83%, Immunology 94%, Biochemistry and Biology 68%, Molecular Biology and Genetics 85%, Neurosciences 72%, Microbiology 76%, Pharmacology/Toxicology 72%). The scientific publishing industry is used for career advancement [36]: publication in specific journals provides scientists with a status signal. As with other luxury items intentionally kept in short supply, there is a motivation to restrict access [37,38].

Some unfavourable consequences may be predicted and some are visible. Resource allocation has long been recognised by economists as problematic in science, especially in basic research where the risks are the greatest. Rival teams undertake unduly dubious and overly similar projects; and too many are attracted to one particular contest to the neglect of other areas, reducing the diversity of areas under exploration [39]. Early decisions by a few influential individuals as to the importance of an area of investigation consolidate path dependency: the first decision predetermines the trajectory. A related effect is herding, where the actions of a few prominent individuals rather than the cumulative input of many independent agents drives people's valuations of a commodity [40,41]. Cascades arise when individuals regard others' earlier actions as more informative than their own private information. The actions upon which people herd may not necessarily be correct; and herding may long continue upon a completely wrong path [41]. Information cascades encourage conventional behaviour, suppress information aggregation, and promote “bubble and bust” cycles. Informational analysis of the literature on molecular interactions in Drosophila genetics has suggested the existence of such information cascades, with positive momentum, interdependence among published papers (most reporting positive data), and dominant themes leading to stagnating conformism [42].

Artificial Scarcity

The authority of journals increasingly derives from their selectivity. The venue of publication provides a valuable status signal. A common excuse for rejection is selectivity based on a limitation ironically irrelevant in the modern age—printed page space. This is essentially an example of artificial scarcity. Artificial scarcity refers to any situation where, even though a commodity exists in abundance, restrictions of access, distribution, or availability make it seem rare, and thus overpriced. Low acceptance rates create an illusion of exclusivity based on merit and more frenzied competition among scientists “selling” manuscripts.

Manuscripts are assessed with a fundamentally negative bias: how they may best be rejected to promote the presumed selectivity of the journal. Journals closely track and advertise their low acceptance rates, equating these with rigorous review: “Nature has space to publish only 10% or so of the 170 papers submitted each week, hence its selection criteria are rigorous”—even though it admits that peer review has a secondary role: “the judgement about which papers will interest a broad readership is made by Nature's editors, not its referees” [43]. Science also equates “high standards of peer review and editorial quality” with the fact that “of the more than 12,000 top-notch scientific manuscripts that the journal sees each year, less than 8% are accepted for publication” [44].

The publication system may operate differently in different fields. For example, for drug trials, journal operations may be dominated by the interests of larger markets: the high volume of transactions involved extends well beyond the small circle of scientific valuations and interests. In other fields where no additional markets are involved (the history of science is perhaps one extreme example), the situation of course may be different. The question to be examined is whether published data may be more representative (and more unbiased) depending on factors such as the ratio of journal outlets to amount of data generated, the relative valuation of specialty journals, career consequences of publication, and accessibility of primary data to the reader.

One solution to artificial scarcity—digital publication—is obvious and already employed. Digital platforms can facilitate the publication of greater numbers of appropriately peer-reviewed manuscripts with reasonable hypotheses and sound methods. Digitally formatted publication need not be limited to few journals, or only to open-access journals. Ideally, all journals could publish in digital form manuscripts that they have received and reviewed and that they consider unsuitable for print publication based on subjective assessments of priority. The current privileging of print over digital publication by some authors and review committees may be reversed, if online-only papers can be demonstrated or perceived to represent equal or better scientific reality than conventional printed manuscripts.


When scientific information itself is the commodity, there is uncertainty as to its value, both immediately and in the long term. Usually we do not know what information will be most useful (valuable) eventually. Economists have struggled with these peculiar attributes of scientific information as a commodity. Production of scientific information is largely paid for by public investment, but the product is offered free to commercial intermediaries, and is culled by them with minimal cost, for sale back to the producers and their underwriters! An explanation for such a strange arrangement is the need for branding—marking the product as valuable. Branding may be more important when a commodity cannot easily be assigned much intrinsic value and when we fear the exchange environment will be flooded with an overabundance of redundant, useless, and misleading product [39,45]. Branding serves a similar and complementary function to the status signal for scientists discussed above. While it is easy to blame journal editors, the industry, or the popular press, there is scant evidence that these actors bear the major culpability [46–49]. Probably authors themselves self-select their work for branding [10,11,50–52].


We may consider several competing or complementary options about the future of scientific publication (Box 1). When economists are asked to analyse a resource-allocation system, a typical assumption is that when information is dispersed, over time, the individual actors will not make systematic errors in their inferences. However, not all economists accept this strong version of rationality. Systematic misperceptions in human behaviour occur with some frequency [53], and misperceptions can perpetuate ineffective systems.

Box 1. Potential Competing or Complementary Options and Solutions for Scientific Publication

  1. Accept the current system as having evolved to be the optimal solution to complex and competing problems.
  2. Promote rapid, digital publication of all articles that contain no flaws, irrespective of perceived “importance”.
  3. Adopt preferred publication of negative over positive results; require very demanding reproducibility criteria before publishing positive results.
  4. Select articles for publication in highly visible venues based on the quality of study methods, their rigorous implementation, and astute interpretation, irrespective of results.
  5. Adopt formal post-publication downward adjustment of claims of papers published in prestigious journals.
  6. Modify current practice to elevate and incorporate more expansive data to accompany print articles or to be accessible in attractive formats associated with high-quality journals: combine the “magazine” and “archive” roles of journals.
  7. Promote critical reviews, digests, and summaries of the large amounts of biomedical data now generated.
  8. Offer disincentives to herding and incentives for truly independent, novel, or heuristic scientific work.
  9. Recognise explicitly and respond to the branding role of journal publication in career development and funding decisions.
  10. Modulate publication practices based on empirical research, which might address correlates of long-term successful outcomes (such as reproducibility, applicability, opening new avenues) of published papers.

Some may accept the current publication system as the ideal culmination of an evolutionary process. However, this order is hardly divinely inspired; additionally, the larger environment has changed over time. Can digital publication alleviate wasteful efforts of repetitive submission, review, revision, and resubmission? Preferred publication of negative over positive results has been suggested, with print publication favoured for all negative data (as more likely to be true) and for only a minority of the positive results that have demonstrated consistency and reproducibility [54]. To exorcise the winner's curse, the quality of experiments rather than the seemingly dramatic results in a minority of them would be the focus of review, but is this feasible in the current reality?

There are limitations to our analysis. Compared with the importance of the problem, there is a relative paucity of empirical observations on the process of scientific publication. The winner's curse is fundamental to our thesis, but there is active debate among economists whether it inhibits real environments or is more of a theoretical phenomenon [1,8]. Do senior investigators make the same adjustments on a high-profile paper's value as do experienced traders on prices? Is herding an appropriate model for scientific publication? Can we correlate the site of publication with the long-term value of individual scientific work and of whole areas of investigation? These questions remain open to analysis and experiment.

Even though its goals may sometimes be usurped for other purposes, science is hard work with limited rewards and only occasional successes. Its interest and importance should speak for themselves, without hyperbole. Uncertainty is powerful and yet quite insufficiently acknowledged when we pretend prescience to guess at the ultimate value of today's endeavours. If “the striving for knowledge and the search for truth are still the strongest motives of scientific discovery”, and if “the advance of science depends upon the free competition of thought” [55], we must ask whether we have created a system for the exchange of scientific ideas that will serve this end.

Supporting Information

Text S1. An extended version of this analytical essay


(115 KB PDF).


Many colleagues have carefully read versions of this work, and the authors express their gratitude in particular to John Barrett, Cynthia Dunbar, Jack Levin, Leonid Margolis, Philip Mortimer, Alan Schechter, Philip Scheinberg, Andrea Young, and Massimo Young.


  1. 1. Thaler RH (1988) Anomalies: The winner's curse. J Econ Perspect 2: 191–202.
  2. 2. Cox JC, Isaac RM (1984) In search of the winner's curse. Econ Inq 22: 579–592.
  3. 3. Dyer D, Kagel JH (1996) Bidding in common value auctions: How the commercial construction industry corrects for the winner's curse. Manage Sci 42: 1463–1475.
  4. 4. Harrison GW, List JA (2007) Naturally occurring markets and exogenous laboratory experiments: A case study of the winner's curse. Cambridge (MA): National Bureau of Economic Research. pp. 1–20. NBER Working Paper No. 13072.
  5. 5. Capen EC, Clapp RV, Campbell WM (1971) Competitive bidding in high-risk situations. J Petrol Technol 23: 641–653.
  6. 6. Cassing J, Douglas RW (1980) Implications of the auction mechanism in baseball's free-agent draft. South Econ J 47: 110–121.
  7. 7. Dessauer JP (1981) Book publishing: What it is, what it does. 2nd edition. New York: R. R. Bowker.
  8. 8. Hendricks K, Porter RH, Boudreau B (1987) Information, returns, and bidding behavior in OCS auctions: 1954–1969. J Ind Econ 35: 517–542.
  9. 9. Ioannidis JPA (2005) Contradicted and initially stronger effects in highly cited clinical research. J Am Med Assoc 294: 218–228.
  10. 10. Krzyzanowska MK, Pintilie M, Tannock IF (2003) Factors associated with failure to publish large randomized trials presented at an oncology meeting. J Am Med Assoc 290: 495–501.
  11. 11. Stern JM (2006) Publication bias: Evidence of delayed publication in a cohort study of clinical research projects. Br Med J 315: 640–645.
  12. 12. Goodman SN (2008) Systematic reviews are not biased by results from trials stopped early for benefit. J Clin Epidemiol 61: 95–96.
  13. 13. Bassler D, Ferreira-Gonzalez I, Briel M, Cook DJ, Devereaux PJ, et al. (2007) Systematic reviewers neglect bias that results from trials stopped early for benefit. J Clin Epidemiol 60: 869–873.
  14. 14. Ioannidis JP (2008) Why most discovered true associations are inflated. Epidemiology 19: 640–648.
  15. 15. Kavvoura FK, Liberopoulos G, Ioannidis JPA (2007) Selection in reported epidemiological risks: An empirical assessment. PLoS Med 4: e79. doi:10.1371/journal.pmed.0040079.
  16. 16. Kyzas PA, Denaxa-Kyza D, Ioannidis JP (2007) Almost all articles on cancer prognostic markers report statistically significant results. Eur J Cancer 43: 2559–2579.
  17. 17. Turner EH, Matthews AM, Linardatos E, Tell RA, Rosenthal R (2008) Selective publication of antidepressant trials and its influence on apparent efficacy. N Eng J Med 358: 252–260.
  18. 18. Ioannidis JPA, Trikalinos TA (2005) Early extreme contradictory estimates may appear in published research: The Proteus phenomenon in molecular genetics research and randomized trials. J Clin Epidemiol 58: 543–549.
  19. 19. Ioannidis JPA (2005) Why most published research findings are false. PLoS Med 2: e124. doi:10.1371/journal.pmed.0020124.
  20. 20. Easterbrook P, Berlin JA, Gopalan R, Matthews DR (1991) Publication bias in clinical research. Lancet 337: 867–872.
  21. 21. Ioannidis JP (2006) Evolution and translation of research findings: From bench to where. PLoS Clin Trials 1: e36. doi:10.1371/journal.pctr.0010036.
  22. 22. Contopoulos-Ioannidis DG, Ntzani E, Ioannidis JP (2003) Translation of highly promising basic science research into clinical applications. Am J Med 114: 477–484.
  23. 23. National Human Genome Research Institute (2008) The Cancer Genome Atlas. Available: Accessed 4 September 2008.
  24. 24. Frank RH, Cook PJ (1995) The winner-take-all society. New York: Free Press.
  25. 25. Goodman SN, Altman DG, George SL (1998) Statistical reviewing policies of medical journals: Caveat lector. J Gen Intern Med 13: 753–756.
  26. 26. Biomed Central (2008) Most viewed articles in past year. Available:​yyear/. Accessed 4 September 2008.
  27. 27. The PLoS Medicine Editors (2006) The impact factor game. PLoS Med 3: e291. doi:10.1371/journal.pmed.0030291.
  28. 28. Smith R (2006) Commentary: The power of the unrelenting impact factor—Is it a force for good or harm. Int J Epidemiol 35: 1129–1130.
  29. 29. Andersen J, Belmont J, Cho CT (2006) Journal impact factor in the era of expanding literature. J Microbiol Immunol Infect 39: 436–443.
  30. 30. Ha TC, Tan SB, Soo KC (2006) The journal impact factor: Too much of an impact. Ann Acad Med Singapore 35: 911–916.
  31. 31. Song F, Eastwood A, Bilbody S, Duley L (1999) The role of electronic journals in reducing publication bias. Med Inform 24: 223–229.
  32. 32. Rossner M, Van Epps H, Hill E (2007) Show me the data. J Exp Med 204: 3052–3053.
  33. 33. Chew M, Villanueva EV, van der Weyden MB (2007) Life and times of the impact factor: Retrospective analysis of trends for seven medical journals (1994–2005) and their editors' views. J Royal Soc Med 100: 142–150.
  34. 34. Ronco C (2006) Scientific journals: Who impacts the impact factor. Int J Artif Organs 29: 645–648.
  35. 35. Ioannidis JPA (2006) Concentration of the most-cited papers in the scientific literature: analysis of journal ecosystems. PLoS ONE 1: e5. doi:10.1371/journal.pone.0000005.
  36. 36. Pan Z, Trikalinos TA, Kavvoura FK, Lau J, Ioannidis JP (2005) Local literature bias in genetic epidemiology: An empirical evaluation of the Chinese literature. PLoS Med 2: e334. doi:10.1371/journal.pmed.0020334.
  37. 37. Ireland NJ (1994) On limiting the market for status signals. J Public Econ 53: 91–110.
  38. 38. Becker GS (1991) A note on restaurant pricing and other examples of social influences on price. Polit Econ 99: 1109–1116.
  39. 39. Dasgupta P, David PA (1994) Toward a new economics of science. Res Pol 23: 487–521.
  40. 40. Bikhchandani S, Hirshleifer D, Welch I (1998) Learning from the behavior of others: Conformity, fads, and informational cascades. J Econ Perspect 12: 151–170.
  41. 41. Hirshleifer D, Teoh SH (2003) Herd behaviour and cascading in capital markets: A review and synthesis. Eur Financ Manage 9: 25–66.
  42. 42. Rzhetsky A, Lossifov I, Loh JM, White KP (2007) Microparadigms: Chains of collective reasoning in publications about molecular interactions. Proc Nat Acad Sci U S A 103: 4940–4945.
  43. 43. Nature (2008) Getting published in Nature: The editorial process. Available:​_published/index.html. Accessed 4 September 2008.
  44. 44. Science (2008) About Science and AAAS. Available:​ut.dtl. Accessed 4 September 2008.
  45. 45. Merton RK (1957) Priorities in scientific discovery: A chapter in the sociology of science. Am Sociol Rev 22: 635–659.
  46. 46. Olson CM, Rennie D, Cook D, Dickersin K, Flanagin A (2002) Publication bias in editorial decision making. J Am Med Assoc 287: 2825–2828.
  47. 47. Brown A, Kraft D, Schmitz SM, Sharpless V, Martin C (2006) Association of industry sponsorship to published outcomes in gastrointestinal clinical research. Clin Gastroenterol Hepatol 4: 1445–1451.
  48. 48. Patsopoulos NA, Analatos AA, Ioannidis JPA (2006) Origin and funding of the most equently cited papers in medicine: Database analysis. Br Med J 332: 1061–1064.
  49. 49. Bubela TM, Caulfield TA (2004) Do the print media “hype” genetic research? A comparison of newspaper stories and peer-reviewed research papers. Canad Med Assoc J 170: 1399–1407.
  50. 50. Polanyi M (1964) Science, faith and society. 2nd edition. Chicago: University of Chicago Press.
  51. 51. Dickersin K, Chan S, Chalmers TC, Sacks HS, Smith H Jr (1987) Publication bias and clinical trials. Control Clin Trials 8: 343–353.
  52. 52. Callaham ML, Wears RL, Weber EJ, Barton C, Young G (1998) Positive-outcome bias and other limitations in the outcome of research abstracts submitted to a scientific meeting. J Am Med Assoc 280: 254–257.
  53. 53. Kahneman D (2003) A perspective on judgment and choice: Mapping bounded rationality. Am Psychol 58: 697–720.
  54. 54. Ioannidis JPA (2006) Journals should publish all “null” results and should sparingly publish “positive” results. Cancer Epidemiol Biomarkers Prev 15: 185.
  55. 55. Popper K (1959) The logic of scientific discovery. New York: Basic Books.