Difference between revisions of "Replication Crisis"

From FreeWiki
Jump to: navigation, search
[checked revision][quality revision]
 
(12 intermediate revisions by the same user not shown)
Line 1: Line 1:
{{Wikipedia1_EN|Replication_Crisis | Replication Crisis|31.8.2019}}
 
  
'''This article is closely connected to the article on the [[Decline Effect]], which is part of this crisis. For further explanation and material please see also there. '''
+
{{Wikipedia2EN|Replication_Crisis | Replication Crisis|31.8.2019}}
  
The '''replication crisis''' (or '''replicability crisis''' or '''reproducibility crisis''') is an ongoing (2019) methodological crisis in which it has been found that many scientific studies are difficult or impossible to replicate or reproduce. The replication crisis affects the social and life sciences most severely.<ref>{{Cite journal | doi = 10.1038/515009a| title = Metascience could rescue the 'replication crisis'| journal = Nature| volume = 515| issue = 7525| pages = 9| year = 2014| last1 = Schooler | first1 = J. W.| pmid=25373639| bibcode = 2014Natur.515....9S}}</ref><ref name="Why 'Statistical Significance' Is Often Insignificant">{{cite web|last1=Smith|first1=Noah|title=Why 'Statistical Significance' Is Often Insignificant|url=https://www.bloomberg.com/view/articles/2017-11-02/why-statistical-significance-is-often-insignificant|website=Bloomberg|accessdate=7 November 2017}}</ref> The crisis has long-standing roots; the phrase was coined in the early 2010s<ref>{{Cite journal | doi = 10.1177/1745691612465253| title = Editors' Introduction to the Special Section on Replicability in Psychological Science: A Crisis of Confidence?| journal = Perspectives on Psychological Science| volume = 7| issue = 6| pages = 528–530| year = 2012| last1=Pashler|first1=Harold |last2=Wagenmakers | first2=Eric Jan| pmid=26168108}}</ref> as part of a growing awareness of the problem. The replication crisis represents an important body of research in [[metascience]].<ref>{{cite web |last1=Fidler |first1=Fiona |last2=Wilcox |first2=John |title=Reproducibility of Scientific Results |url=https://plato.stanford.edu/entries/scientific-reproducibility/#MetaScieEstaMoniEvalReprCris |website=The Stanford Encyclopedia of Philosophy |publisher=Metaphysics Research Lab, Stanford University |accessdate=19 May 2019 |date=2018}}</ref>
+
'''This article is closely connected to the article on the [[Decline_Effect|Decline Effect]], which is part of this crisis. For further explanation and material please see also there.'''
  
Because the reproducibility of experiments is an essential part of the scientific method,<ref>Staddon, John (2017) Scientific Method: How science works, fails to work or pretends to work.  Taylor and Francis.</ref> the inability to replicate the studies of others has potentially grave consequences for many fields of science in which significant theories are grounded on unreproducible experimental work. The replication crisis has been particularly widely discussed in the field of psychology (and in particular, social psychology) and in [[medicine]], where a number of efforts have been made to re-investigate classic results, and to attempt to determine both the reliability of the results, and, if found to be unreliable, the reasons for the failure of replication.<ref>{{cite web|url=http://www.newyorker.com/tech/elements/the-crisis-in-social-psychology-that-isnt|title=The Crisis in Social Psychology That Isn't|author=Gary Marcus|publisher=The New Yorker|date=May 1, 2013}}</ref><ref>{{cite web|url=http://www.newyorker.com/magazine/2010/12/13/the-truth-wears-off|title=The Truth Wears Off|author=Jonah Lehrer|publisher=The New Yorker|date=December 13, 2010}}</ref>
+
The '''replication crisis''' (or '''replicability crisis''' or '''reproducibility crisis''') is an ongoing (2019) methodological crisis in which it has been found that many scientific studies are difficult or impossible to replicate or reproduce. The replication crisis affects the social and life sciences most severely,<ref>{{Cite journal | doi = 10.1038/515009a| title = Metascience could rescue the 'replication crisis'| journal = Nature| volume = 515| issue = 7525| pages = 9| year = 2014| last1 = Schooler | first1 = J. W.| pmid=25373639| bibcode = 2014Natur.515....9S}}</ref><ref name="Why 'Statistical Significance' Is Often Insignificant">{{cite web|last1=Smith|first1=Noah|title=Why 'Statistical Significance' Is Often Insignificant|url=https://www.bloomberg.com/view/articles/2017-11-02/why-statistical-significance-is-often-insignificant|website=Bloomberg|accessdate=7 November 2017}}</ref> whereas the hard sciences are still reluctant to face the problem in the first place and mainly don’t even use double blinding in their research.<ref>Rupert Sheldrake: The Science Delusion – Freeing the Spirit of Enquiry; London 2012, Hodder& Stoughton, ISBN 978 1 444 72795 1. Chapter 11. Illusions of Objectivity.</ref> The crisis has long-standing roots. The phrase was coined in the early 2010s<ref>{{Cite journal | doi = 10.1177/1745691612465253| title = Editors' Introduction to the Special Section on Replicability in Psychological Science: A Crisis of Confidence?| journal = Perspectives on Psychological Science| volume = 7| issue = 6| pages = 528–530| year = 2012| last1=Pashler|first1=Harold |last2=Wagenmakers | first2=Eric Jan| pmid=26168108}}</ref> as part of a growing awareness of the problem. The replication crisis represents an important body of research in metascience.<ref>{{cite web |last1=Fidler |first1=Fiona |last2=Wilcox |first2=John |title=Reproducibility of Scientific Results |url=https://plato.stanford.edu/entries/scientific-reproducibility/#MetaScieEstaMoniEvalReprCris |website=The Stanford Encyclopedia of Philosophy |publisher=Metaphysics Research Lab, Stanford University |accessdate=19 May 2019 |date=2018}}</ref>
 +
 
 +
Because the reproducibility of experiments is an essential part of the scientific method,<ref>Staddon, John (2017) Scientific Method: How science works, fails to work or pretends to work.  Taylor and Francis.</ref> the inability to replicate the studies of others has potentially grave consequences for many fields of science in which significant theories are grounded on unreproducible experimental work. The replication crisis has been particularly widely discussed in the field of psychology (and in particular, social psychology) and in medicine, where a number of efforts have been made to re-investigate classic results, and to attempt to determine both the reliability of the results, and, if found to be unreliable, the reasons for the failure of replication.<ref>{{cite web|url=http://www.newyorker.com/tech/elements/the-crisis-in-social-psychology-that-isnt|title=The Crisis in Social Psychology That Isn't|author=Gary Marcus|publisher=The New Yorker|date=May 1, 2013}}</ref><ref>{{cite web|url=http://www.newyorker.com/magazine/2010/12/13/the-truth-wears-off|title=The Truth Wears Off|author=Jonah Lehrer|publisher=The New Yorker|date=December 13, 2010}}</ref>
 +
 
 +
== Scope of the crisis ==
 +
 
 +
=== Overall ===
  
==Scope of the crisis==
 
===Overall===
 
 
According to a 2016 poll of 1,500 scientists reported that 70% of them had failed to reproduce at least one other scientist's experiment (50% had failed to reproduce one of their own experiments).<ref>{{cite web|title=Is There a Reproducibility Crisis in Science?|url=https://www.scientificamerican.com/video/is-there-a-reproducibility-crisis-in-science/|publisher=Nature Video, Scientific American|date=28 May 2016|accessdate=15 August 2019}}</ref> In 2009, 2% of scientists admitted to falsifying studies at least once and 14% admitted to personally knowing someone who did. Misconducts were reported more frequently by medical researchers than others.<ref>{{cite journal|title=How Many Scientists Fabricate and Falsify Research? A Systematic Review and Meta-Analysis of Survey Data|first=Daniele|last=Fanelli|date=29 May 2009|journal=PLOS ONE|volume=4|issue=5|pages=e5738|doi=10.1371/journal.pone.0005738|pmid=19478950|pmc=2685008|bibcode=2009PLoSO...4.5738F}}</ref>
 
According to a 2016 poll of 1,500 scientists reported that 70% of them had failed to reproduce at least one other scientist's experiment (50% had failed to reproduce one of their own experiments).<ref>{{cite web|title=Is There a Reproducibility Crisis in Science?|url=https://www.scientificamerican.com/video/is-there-a-reproducibility-crisis-in-science/|publisher=Nature Video, Scientific American|date=28 May 2016|accessdate=15 August 2019}}</ref> In 2009, 2% of scientists admitted to falsifying studies at least once and 14% admitted to personally knowing someone who did. Misconducts were reported more frequently by medical researchers than others.<ref>{{cite journal|title=How Many Scientists Fabricate and Falsify Research? A Systematic Review and Meta-Analysis of Survey Data|first=Daniele|last=Fanelli|date=29 May 2009|journal=PLOS ONE|volume=4|issue=5|pages=e5738|doi=10.1371/journal.pone.0005738|pmid=19478950|pmc=2685008|bibcode=2009PLoSO...4.5738F}}</ref>
  
=== In psychology ===
+
=== Psychology ===
Several factors have combined to put psychology at the center of the controversy.<ref>{{cite web|last1=Achenbach|first1=Joel|title=No, science's reproducibility problem is not limited to psychology|url=https://www.washingtonpost.com/news/speaking-of-science/wp/2015/08/28/no-sciences-reproducibility-problem-is-not-limited-to-psychology/|website=The Washington Post|accessdate=10 September 2015}}</ref> Much of the focus has been on the area of [[social psychology]],<ref>{{Cite news|url=https://www.nytimes.com/2017/10/18/magazine/when-the-revolution-came-for-amy-cuddy.html|title=When the Revolution Came for Amy Cuddy|last=Dominus|first=Susan|date=2017-10-18|work=The New York Times|access-date=2017-10-19|language=en-US|issn=0362-4331}}</ref> although other areas of psychology such as [[clinical psychology]],<ref>{{cite journal|last1=Leichsenring|first1=Falk|last2=Abbass|first2=Allan|author-link2=Allan Abbass|last3=Hilsenroth|first3=Mark J.|last4=Leweke|first4=Frank|last5=Luyten|first5=Patrick|last6=Keefe|first6=Jack R.|last7=Midgley|first7=Nick|last8=Rabung|first8=Sven|last9=Salzer|first9=Simone|last10=Steiner|first10=Christiane|title=Biases in research: risk factors for non-replicability in psychotherapy and pharmacotherapy research|journal=[[Psychological Medicine]]|year=April 2017|volume=47|issue=6|pages=1000–1011|doi=10.1017/S003329171600324X|pmid=27955715|url=http://discovery.ucl.ac.uk/1532689/}}</ref><ref>{{cite journal|last=Hengartner|first=Michael P.|title=Raising Awareness for the Replication Crisis in Clinical Psychology by Focusing on Inconsistencies in Psychotherapy Research: How Much Can We Rely on Published Findings from Efficacy Trials?|journal=[[Frontiers in Psychology]]|volume=9|pages=256|publisher=[[Frontiers Media]]|doi=10.3389/fpsyg.2018.00256|pmid=29541051|pmc=5835722|date=February 28, 2018}}</ref> [[developmental psychology]],<ref>{{cite journal|last1=Frank|first1=Michael C.|author-link1=Michael C. Frank|last2=Bergelson|first2=Elika|last3=Bergmann|first3=Christina|last4=Cristia|first4=Alejandrina|last5=Floccia|first5=Caroline|last6=Gervain|first6=Judit|last7=Hamlin|first7=J. Kiley|last8=Hannon|first8=Erin E.|last9=Kline|first9=Melissa|last10=Levelt|first10=Claartje|last11=Lew-Williams|first11=Casey|last12=Nazzi|first12=Thierry|last13=Panneton|first13=Robin|last14=Rabagliati|first14=Hugh|last15=Soderstrom|first15=Melanie|last16=Sullivan|first16=Jessica|last17=Waxman|first17=Sandra|author-link17=Sandra Waxman|last18=Yurovsky|first18=Daniel|title=A Collaborative Approach to Infant Research: Promoting Reproducibility, Best Practices, and Theory‐Building|journal=[[Infancy (journal)|Infancy]]|volume=22|issue=4|pages=421–435|date=9 March 2017|url=http://callab.uchicago.edu/papers/fetal-infancy-inpress.pdf|doi=10.1111/infa.12182|access-date=19 December 2018|hdl=10026.1/9942}}</ref> and [[educational research]] have also been implicated.<ref>{{cite journal|last=Tyson|first=Charlie|title=Failure to Replicate|journal=[[Inside Higher Ed]]|date=14 August 2014|url=https://www.insidehighered.com/news/2014/08/14/almost-no-education-research-replicated-new-article-shows|access-date=19 December 2018}}</ref><ref>{{cite journal|last1=Makel|first1=Matthew C.|last2=Plucker|first2=Jonathan A.|author-link2=Jonathan A. Plucker|title=Facts Are More Important Than Novelty: Replication in the Education Sciences|journal=[[Educational Researcher]]|date=1 August 2014|volume=43|issue=6|pages=304–316|doi=10.3102/0013189X14545513|access-date=19 December 2018|url=https://journals.sagepub.com/stoken/rbtfl/w5mrNxPVD8zSg/full}}</ref> According to a 2018 survey of 200 meta-analyses, "psychological research is, on average, afflicted with low statistical power".<ref>{{Cite journal|last=Stanley|first=T. D.|last2=Carter|first2=Evan C.|last3=Doucouliagos|first3=Hristos|date=2018|title=What meta-analyses reveal about the replicability of psychological research.|journal=Psychological Bulletin|language=en|volume=144|issue=12|pages=1325–1346|doi=10.1037/bul0000169|pmid=30321017|issn=1939-1455}}</ref>
+
 
 +
Several factors have combined to put psychology at the center of the controversy.<ref>{{cite web|last1=Achenbach|first1=Joel|title=No, science's reproducibility problem is not limited to psychology|url=https://www.washingtonpost.com/news/speaking-of-science/wp/2015/08/28/no-sciences-reproducibility-problem-is-not-limited-to-psychology/|website=The Washington Post|accessdate=10 September 2015}}</ref> Much of the focus has been on the area of social psychology,<ref>{{Cite news|url=https://www.nytimes.com/2017/10/18/magazine/when-the-revolution-came-for-amy-cuddy.html|title=When the Revolution Came for Amy Cuddy|last=Dominus|first=Susan|date=2017-10-18|work=The New York Times|access-date=2017-10-19|language=en-US|issn=0362-4331}}</ref> although other areas of psychology such as clinical psychology,<ref>{{cite journal|last1=Leichsenring|first1=Falk|last2=Abbass|first2=Allan|author-link2=Allan Abbass|last3=Hilsenroth|first3=Mark J.|last4=Leweke|first4=Frank|last5=Luyten|first5=Patrick|last6=Keefe|first6=Jack R.|last7=Midgley|first7=Nick|last8=Rabung|first8=Sven|last9=Salzer|first9=Simone|last10=Steiner|first10=Christiane|title=Biases in research: risk factors for non-replicability in psychotherapy and pharmacotherapy research|journal=[[Psychological Medicine]]|year=April 2017|volume=47|issue=6|pages=1000–1011|doi=10.1017/S003329171600324X|pmid=27955715|url=http://discovery.ucl.ac.uk/1532689/}}</ref><ref>{{cite journal|last=Hengartner|first=Michael P.|title=Raising Awareness for the Replication Crisis in Clinical Psychology by Focusing on Inconsistencies in Psychotherapy Research: How Much Can We Rely on Published Findings from Efficacy Trials?|journal=[[Frontiers in Psychology]]|volume=9|pages=256|publisher=[[Frontiers Media]]|doi=10.3389/fpsyg.2018.00256|pmid=29541051|pmc=5835722|date=February 28, 2018}}</ref> developmental psychology,<ref>{{cite journal|last1=Frank|first1=Michael C.|author-link1=Michael C. Frank|last2=Bergelson|first2=Elika|last3=Bergmann|first3=Christina|last4=Cristia|first4=Alejandrina|last5=Floccia|first5=Caroline|last6=Gervain|first6=Judit|last7=Hamlin|first7=J. Kiley|last8=Hannon|first8=Erin E.|last9=Kline|first9=Melissa|last10=Levelt|first10=Claartje|last11=Lew-Williams|first11=Casey|last12=Nazzi|first12=Thierry|last13=Panneton|first13=Robin|last14=Rabagliati|first14=Hugh|last15=Soderstrom|first15=Melanie|last16=Sullivan|first16=Jessica|last17=Waxman|first17=Sandra|author-link17=Sandra Waxman|last18=Yurovsky|first18=Daniel|title=A Collaborative Approach to Infant Research: Promoting Reproducibility, Best Practices, and Theory‐Building|journal=[[Infancy (journal)|Infancy]]|volume=22|issue=4|pages=421–435|date=9 March 2017|url=http://callab.uchicago.edu/papers/fetal-infancy-inpress.pdf|doi=10.1111/infa.12182|access-date=19 December 2018|hdl=10026.1/9942}}</ref> and educational research have also been implicated.<ref>{{cite journal|last=Tyson|first=Charlie|title=Failure to Replicate|journal=[[Inside Higher Ed]]|date=14 August 2014|url=https://www.insidehighered.com/news/2014/08/14/almost-no-education-research-replicated-new-article-shows|access-date=19 December 2018}}</ref><ref>{{cite journal|last1=Makel|first1=Matthew C.|last2=Plucker|first2=Jonathan A.|author-link2=Jonathan A. Plucker|title=Facts Are More Important Than Novelty: Replication in the Education Sciences|journal=[[Educational Researcher]]|date=1 August 2014|volume=43|issue=6|pages=304–316|doi=10.3102/0013189X14545513|access-date=19 December 2018|url=https://journals.sagepub.com/stoken/rbtfl/w5mrNxPVD8zSg/full}}</ref> According to a 2018 survey of 200 meta-analyses, "psychological research is, on average, afflicted with low statistical power".<ref>{{Cite journal|last=Stanley|first=T. D.|last2=Carter|first2=Evan C.|last3=Doucouliagos|first3=Hristos|date=2018|title=What meta-analyses reveal about the replicability of psychological research.|journal=Psychological Bulletin|language=en|volume=144|issue=12|pages=1325–1346|doi=10.1037/bul0000169|pmid=30321017|issn=1939-1455}}</ref>
  
Firstly, questionable research practices (QRPs) have been identified as common in the field.<ref name=":0"/> Such practices, while not intentionally fraudulent, involve capitalizing on the gray area of acceptable scientific practices or exploiting flexibility in data collection, analysis, and reporting, often in an effort to obtain a desired outcome. Examples of QRPs include selective reporting or partial publication of data (reporting only some of the study conditions or collected dependent measures in a publication), optional stopping (choosing when to stop data collection, often based on statistical significance of tests), [[p-value|''p''-value]] rounding (rounding ''p''-values down to 0.05 to suggest statistical significance), file drawer effect (nonpublication of data), post-hoc storytelling (framing exploratory analyses as confirmatory analyses), and manipulation of outliers (either removing outliers or leaving outliers in a dataset to cause a statistical test to be significant).<ref name=":0">{{Cite journal|title = Measuring the Prevalence of Questionable Research Practices With Incentives for Truth Telling|url = https://www.hbs.edu/faculty/Publication%20Files/QRP12_a137cc09-45bc-4eae-9ee9-c0cb47f14add.pdf|journal = Psychological Science|date = 2012-05-01|issn = 0956-7976|pmid = 22508865|pages = 524–532|volume = 23|issue = 5|doi = 10.1177/0956797611430953|first = Leslie K.|last = John|first2 = George|last2 = Loewenstein|first3 = Drazen|last3 = Prelec}}</ref><ref>{{Cite journal|title = The Nine Circles of Scientific Hell|journal = Perspectives on Psychological Science|date = 2012-11-01|issn = 1745-6916|pmid = 26168124|pages = 643–644|volume = 7|issue = 6|doi = 10.1177/1745691612459519}}</ref><ref>{{Cite web|title = Research misconduct - The grey area of Questionable Research Practices|url = http://www.vib.be/en/news/Pages/Research-misconduct---The-grey-area-of-Questionable-Research-Practices.aspx|website = www.vib.be|accessdate = 2015-11-13}}</ref><ref>{{Cite journal|title = Questionable Research Practices Revisited|journal = Social Psychological and Personality Science|date = 2015-10-19|issn = 1948-5506|pages = 45–52|doi = 10.1177/1948550615612150|first = Klaus|last = Fiedler|first2 = Norbert|last2 = Schwarz|volume=7}}</ref> A survey of over 2,000 psychologists indicated that a majority of respondents admitted to using at least one QRP.<ref name=":0" /> [[False positive]] conclusions, often resulting from the [[Publish or perish|pressure to publish]] or the author's own [[confirmation bias]], are an inherent hazard in the field, requiring a certain degree of [[skepticism]] on the part of readers.<ref name="Simmons et al. (2011)">{{cite journal|last1 = Simmons|first1 = Joseph|first2 = Leif|last2 = Nelson|first3 = Uri|last3 = Simonsohn|title = False-Positive Psychology: Undisclosed Flexibility in Data Collection and Analysis Allows Presenting Anything as Significant|journal = [[Psychological Science]]|volume = 22|issue = 11|pages = 1359–1366|date = November 2011|issn = 0956-7976|doi = 10.1177/0956797611417632|pmid = 22006061}}</ref>
+
Firstly, questionable research practices (QRPs) have been identified as common in the field.<ref name=":0" /> Such practices, while not intentionally fraudulent, involve capitalizing on the gray area of acceptable scientific practices or exploiting flexibility in data collection, analysis, and reporting, often in an effort to obtain a desired outcome. Examples of QRPs include selective reporting or partial publication of data (reporting only some of the study conditions or collected dependent measures in a publication), optional stopping (choosing when to stop data collection, often based on statistical significance of tests), ''p''-value rounding (rounding ''p''-values down to 0.05 to suggest statistical significance), file drawer effect (nonpublication of data), post-hoc storytelling (framing exploratory analyses as confirmatory analyses), and manipulation of outliers (either removing outliers or leaving outliers in a dataset to cause a statistical test to be significant).<ref name=":0">{{Cite journal|title = Measuring the Prevalence of Questionable Research Practices With Incentives for Truth Telling|url = https://www.hbs.edu/faculty/Publication%20Files/QRP12_a137cc09-45bc-4eae-9ee9-c0cb47f14add.pdf|journal = Psychological Science|date = 2012-05-01|issn = 0956-7976|pmid = 22508865|pages = 524–532|volume = 23|issue = 5|doi = 10.1177/0956797611430953|first = Leslie K.|last = John|first2 = George|last2 = Loewenstein|first3 = Drazen|last3 = Prelec}}</ref><ref>{{Cite journal|title = The Nine Circles of Scientific Hell|journal = Perspectives on Psychological Science|date = 2012-11-01|issn = 1745-6916|pmid = 26168124|pages = 643–644|volume = 7|issue = 6|doi = 10.1177/1745691612459519}}</ref><ref>{{Cite web|title = Research misconduct - The grey area of Questionable Research Practices|url = http://www.vib.be/en/news/Pages/Research-misconduct---The-grey-area-of-Questionable-Research-Practices.aspx|website = www.vib.be|accessdate = 2015-11-13}}</ref><ref>{{Cite journal|title = Questionable Research Practices Revisited|journal = Social Psychological and Personality Science|date = 2015-10-19|issn = 1948-5506|pages = 45–52|doi = 10.1177/1948550615612150|first = Klaus|last = Fiedler|first2 = Norbert|last2 = Schwarz|volume=7}}</ref> A survey of over 2,000 psychologists indicated that a majority of respondents admitted to using at least one QRP.<ref name=":0" /> False positive conclusions, often resulting from the pressure to publish or the author's own confirmation bias, are an inherent hazard in the field, requiring a certain degree of skepticism on the part of readers.<ref name="Simmons et al. (2011)">{{cite journal|last1 = Simmons|first1 = Joseph|first2 = Leif|last2 = Nelson|first3 = Uri|last3 = Simonsohn|title = False-Positive Psychology: Undisclosed Flexibility in Data Collection and Analysis Allows Presenting Anything as Significant|journal = [[Psychological Science]]|volume = 22|issue = 11|pages = 1359–1366|date = November 2011|issn = 0956-7976|doi = 10.1177/0956797611417632|pmid = 22006061}}</ref>
  
Secondly, psychology and social psychology in particular, has found itself at the center of several scandals involving outright fraudulent research, most notably the admitted data fabrication by [[Diederik Stapel]]<ref>{{cite journal|url=http://chronicle.com/article/As-Dutch-Research-Scandal/129746/|title=Fraud Scandal Fuels Debate Over Practices of Social Psychology|journal=The Chronicle of Higher Education|first=Christopher|last=Shea|date=13 November 2011}}</ref> as well as allegations against others. However, most scholars acknowledge that fraud is, perhaps, the lesser contribution to replication crises.
+
Secondly, psychology and social psychology in particular, has found itself at the center of several scandals involving outright fraudulent research, most notably the admitted data fabrication by Diederik Stapel<ref>{{cite journal|url=http://chronicle.com/article/As-Dutch-Research-Scandal/129746/|title=Fraud Scandal Fuels Debate Over Practices of Social Psychology|journal=The Chronicle of Higher Education|first=Christopher|last=Shea|date=13 November 2011}}</ref> as well as allegations against others. However, most scholars acknowledge that fraud is, perhaps, the lesser contribution to replication crises.
  
Thirdly, several effects in psychological science have been found to be difficult to replicate even before the current replication crisis. For example, the scientific journal ''[[Society for Judgment and Decision Making#Publications and journal|Judgment and Decision Making]]'' has published several studies over the years that fail to provide support for the [[unconscious thought theory]]. Replications appear particularly difficult when research trials are pre-registered and conducted by research groups not highly invested in the theory under questioning.
+
Thirdly, several effects in psychological science have been found to be difficult to replicate even before the current replication crisis. For example, the scientific journal ''Judgment and Decision Making'' has published several studies over the years that fail to provide support for the unconscious thought theory. Replications appear particularly difficult when research trials are pre-registered and conducted by research groups not highly invested in the theory under questioning.
  
These three elements together have resulted in renewed attention for replication supported by psychologist [[Daniel Kahneman]].<ref>{{cite web|title=A New Etiquette for Replication|url=https://www.scribd.com/doc/225285909/Kahneman-Commentary|first=Daniel|last=Kahneman|via=Scribd}}</ref> Scrutiny of many effects have shown that several core beliefs are hard to replicate. A recent special edition of the journal ''Social Psychology'' focused on replication studies and a number of previously held beliefs were found to be difficult to replicate.<ref>{{cite journal|year=2014 |title= Link to issue<!-- NONE - reference to entire issue -->|url=http://www.psycontent.com/content/l67413865317/?p=3f8268e0307f4036906c158b1667c57a&pi=0 |journal=Social Psychology |publisher=Hogrefe Publishing |volume=45 |issue=3 |archive-url=https://archive.is/20140530230728/http://www.psycontent.com/content/l67413865317/?p=3f8268e0307f4036906c158b1667c57a&pi=0 |archive-date=30 May 2014}}</ref> A 2012 special edition of the journal ''Perspectives on Psychological Science'' also focused on issues ranging from publication bias to null-aversion that contribute to the replication crises in psychology.<ref>[http://pps.sagepub.com/content/7/6.toc Table of Contents]</ref> In 2015, the first open empirical study of reproducibility in Psychology was published, called the [[Reproducibility Project]]. Researchers from around the world collaborated to replicate 100 empirical studies from three top Psychology journals. Fewer than half of the attempted replications were successful at producing statistically significant results in the expected directions, though most of the attempted replications did produce trends in the expected directions.<ref>{{Cite journal | last1 = Open Science Collaboration | title = Estimating the reproducibility of Psychological Science | doi = 10.1126/science.aac4716 | journal = Science | volume = 349 | issue = 6251 | pages = aac4716 | year = 2015 | pmid=26315443| url = http://eprints.keele.ac.uk/877/1/Open%20Science%20%28Science%20Pre-Print%29.pdf | hdl = 10722/230596 }}</ref>
+
These three elements together have resulted in renewed attention for replication supported by psychologist Daniel Kahneman.<ref>{{cite web|title=A New Etiquette for Replication|url=https://www.scribd.com/doc/225285909/Kahneman-Commentary|first=Daniel|last=Kahneman|via=Scribd}}</ref> Scrutiny of many effects have shown that several core beliefs are hard to replicate. A recent special edition of the journal ''Social Psychology'' focused on replication studies and a number of previously held beliefs were found to be difficult to replicate.<ref>{{cite journal|year=2014 |title= Link to issue<!-- NONE - reference to entire issue -->|url=http://www.psycontent.com/content/l67413865317/?p=3f8268e0307f4036906c158b1667c57a&pi=0 |journal=Social Psychology |publisher=Hogrefe Publishing |volume=45 |issue=3 |archive-url=https://archive.is/20140530230728/http://www.psycontent.com/content/l67413865317/?p=3f8268e0307f4036906c158b1667c57a&pi=0 |archive-date=30 May 2014}}</ref> A 2012 special edition of the journal ''Perspectives on Psychological Science'' also focused on issues ranging from publication bias to null-aversion that contribute to the replication crises in psychology.<ref>[http://pps.sagepub.com/content/7/6.toc Table of Contents]</ref> In 2015, the first open empirical study of reproducibility in Psychology was published, called the Reproducibility Project. Researchers from around the world collaborated to replicate 100 empirical studies from three top Psychology journals. Fewer than half of the attempted replications were successful at producing statistically significant results in the expected directions, though most of the attempted replications did produce trends in the expected directions.<ref>{{Cite journal | last1 = Open Science Collaboration | title = Estimating the reproducibility of Psychological Science | doi = 10.1126/science.aac4716 | journal = Science | volume = 349 | issue = 6251 | pages = aac4716 | year = 2015 | pmid=26315443| url = http://eprints.keele.ac.uk/877/1/Open%20Science%20%28Science%20Pre-Print%29.pdf | hdl = 10722/230596 }}</ref>
  
 
Many research trials and meta-analyses are compromised by poor quality and conflicts of interest that involve both authors and professional advocacy organizations, resulting in many false positives regarding the effectiveness of certain types of psychotherapy.<ref>{{cite web | url=http://blogs.plos.org/mindthebrain/2014/04/15/meta-analyses-conducted-professional-organizations-trustworthy/ | title=Are meta analyses conducted by professional organizations more trustworthy? | work=Mind the Brain | date=April 15, 2014 | accessdate=September 13, 2016 | first=James | last=Coyne}}</ref>
 
Many research trials and meta-analyses are compromised by poor quality and conflicts of interest that involve both authors and professional advocacy organizations, resulting in many false positives regarding the effectiveness of certain types of psychotherapy.<ref>{{cite web | url=http://blogs.plos.org/mindthebrain/2014/04/15/meta-analyses-conducted-professional-organizations-trustworthy/ | title=Are meta analyses conducted by professional organizations more trustworthy? | work=Mind the Brain | date=April 15, 2014 | accessdate=September 13, 2016 | first=James | last=Coyne}}</ref>
  
Although the British newspaper ''[[The Independent]]'' wrote that the results of the [[reproducibility project]] show that much of the published research is just "psycho-babble",<ref>{{cite news| url=https://www.independent.co.uk/news/science/study-reveals-that-a-lot-of-psychology-research-really-is-just-psychobabble-10474646.html | location=London | work=The Independent | first=Steve | last=Connor | title=Study reveals that a lot of psychology research really is just 'psycho-babble' | date=27 August 2015}}</ref> the replication crisis does not necessarily mean that psychology is unscientific.<ref>{{cite web|url=http://www.slate.com/articles/health_and_science/science/2014/07/replication_controversy_in_psychology_bullying_file_drawer_effect_blog_posts.single.html|title=Why Psychologists' Food Fight Matters|first1=Michelle N.|last1=Meyer|first2=Christopher|last2=Chabris|date=31 July 2014|via=Slate}}</ref><ref>{{cite web|url=http://fivethirtyeight.com/datalab/psychology-is-starting-to-deal-with-its-replication-problem/|title=Psychology Is Starting To Deal With Its Replication Problem|date=27 August 2015}}</ref><ref>{{cite web|url=http://fivethirtyeight.com/features/science-isnt-broken/|title=Science Isn't Broken|date=19 August 2015}}</ref> Rather this process is a healthy if sometimes acrimonious part of the scientific process in which old ideas or those that cannot withstand careful scrutiny are pruned,<ref>{{cite news|url=https://www.theguardian.com/science/head-quarters/2014/may/28/psychology-replication-drive-methods-bullying|title=Psychology's replication drive: it's not about you|first=Pete|last=Etchells|date=28 May 2014|newspaper=The Guardian}}</ref><ref>{{Cite journal|title = An Agenda for Purely Confirmatory Research|journal = Perspectives on Psychological Science|date = 2012-11-01|issn = 1745-6916|pmid = 26168122|pages = 632–638|volume = 7|issue = 6|doi = 10.1177/1745691612463078|first = Eric-Jan|last = Wagenmakers|first2 = Ruud|last2 = Wetzels|first3 = Denny|last3 = Borsboom|first4 = Han L. J. van der|last4 = Maas|first5 = Rogier A.|last5 = Kievit}}</ref> although this pruning process is not always effective.<ref>{{Cite journal|title = Why Science Is Not Necessarily Self-Correcting|journal = Perspectives on Psychological Science|date = 2012-11-01|issn = 1745-6916|pmid = 26168125|pages = 645–654|volume = 7|issue = 6|doi = 10.1177/1745691612464056|first = John P. A.|last = Ioannidis}}</ref><ref>{{Cite journal|title = Is the Replicability Crisis Overblown? Three Arguments Examined|journal = Perspectives on Psychological Science|date = 2012-11-01|issn = 1745-6916|pmid = 26168109|pages = 531–536|volume = 7|issue = 6|doi = 10.1177/1745691612463401|first = Harold|last = Pashler|first2 = Christine R.|last2 = Harris}}</ref> The consequence is that some areas of psychology once considered solid, such as social priming, have come under increased scrutiny due to failed replications.<ref>{{cite journal|url=http://chronicle.com/article/Power-of-Suggestion/136907/|title=Power of Suggestion|journal=The Chronicle of Higher Education|first=Tom|last=Bartlett|date=30 January 2013}}</ref>
+
Although the British newspaper ''The Independent'' wrote that the results of the reproducibility project show that much of the published research is just "psycho-babble",<ref>{{cite news| url=https://www.independent.co.uk/news/science/study-reveals-that-a-lot-of-psychology-research-really-is-just-psychobabble-10474646.html | location=London | work=The Independent | first=Steve | last=Connor | title=Study reveals that a lot of psychology research really is just 'psycho-babble' | date=27 August 2015}}</ref> the replication crisis does not necessarily mean that psychology is unscientific.<ref>{{cite web|url=http://www.slate.com/articles/health_and_science/science/2014/07/replication_controversy_in_psychology_bullying_file_drawer_effect_blog_posts.single.html|title=Why Psychologists' Food Fight Matters|first1=Michelle N.|last1=Meyer|first2=Christopher|last2=Chabris|date=31 July 2014|via=Slate}}</ref><ref>{{cite web|url=http://fivethirtyeight.com/datalab/psychology-is-starting-to-deal-with-its-replication-problem/|title=Psychology Is Starting To Deal With Its Replication Problem|date=27 August 2015}}</ref><ref>{{cite web|url=http://fivethirtyeight.com/features/science-isnt-broken/|title=Science Isn't Broken|date=19 August 2015}}</ref> Rather this process is a healthy if sometimes acrimonious part of the scientific process in which old ideas or those that cannot withstand careful scrutiny are pruned,<ref>{{cite news|url=https://www.theguardian.com/science/head-quarters/2014/may/28/psychology-replication-drive-methods-bullying|title=Psychology's replication drive: it's not about you|first=Pete|last=Etchells|date=28 May 2014|newspaper=The Guardian}}</ref><ref>{{Cite journal|title = An Agenda for Purely Confirmatory Research|journal = Perspectives on Psychological Science|date = 2012-11-01|issn = 1745-6916|pmid = 26168122|pages = 632–638|volume = 7|issue = 6|doi = 10.1177/1745691612463078|first = Eric-Jan|last = Wagenmakers|first2 = Ruud|last2 = Wetzels|first3 = Denny|last3 = Borsboom|first4 = Han L. J. van der|last4 = Maas|first5 = Rogier A.|last5 = Kievit}}</ref> although this pruning process is not always effective.<ref>{{Cite journal|title = Why Science Is Not Necessarily Self-Correcting|journal = Perspectives on Psychological Science|date = 2012-11-01|issn = 1745-6916|pmid = 26168125|pages = 645–654|volume = 7|issue = 6|doi = 10.1177/1745691612464056|first = John P. A.|last = Ioannidis}}</ref><ref>{{Cite journal|title = Is the Replicability Crisis Overblown? Three Arguments Examined|journal = Perspectives on Psychological Science|date = 2012-11-01|issn = 1745-6916|pmid = 26168109|pages = 531–536|volume = 7|issue = 6|doi = 10.1177/1745691612463401|first = Harold|last = Pashler|first2 = Christine R.|last2 = Harris}}</ref> The consequence is that some areas of psychology once considered solid, such as social priming, have come under increased scrutiny due to failed replications.<ref>{{cite journal|url=http://chronicle.com/article/Power-of-Suggestion/136907/|title=Power of Suggestion|journal=The Chronicle of Higher Education|first=Tom|last=Bartlett|date=30 January 2013}}</ref>
  
Nobel laureate and professor emeritus in psychology Daniel Kahneman argued that the original authors should be involved in the replication effort because the published methods are often too vague.<ref name="Chambers">{{cite news|url=https://www.theguardian.com/science/head-quarters/2014/jun/10/physics-envy-do-hard-sciences-hold-the-solution-to-the-replication-crisis-in-psychology|title=Physics envy: Do 'hard' sciences hold the solution to the replication crisis in psychology?|first=Chris|last=Chambers|date=10 June 2014|newspaper=The Guardian}}</ref> Others such as Dr. Andrew Wilson disagree and argue that the methods should be written down in detail.<ref name="Chambers"/> An investigation of replication rates in psychology in 2012 indicated higher success rates of replication in replication studies when there was author overlap with the original authors of a study<ref name=":1">{{Cite journal|title = Replications in Psychology Research How Often Do They Really Occur?|journal = Perspectives on Psychological Science|date = 2012-11-01|issn = 1745-6916|pmid = 26168110|pages = 537–542|volume = 7|issue = 6|doi = 10.1177/1745691612460688|first = Matthew C.|last = Makel|first2 = Jonathan A.|last2 = Plucker|first3 = Boyd|last3 = Hegarty}}</ref> (91.7% successful replication rates in studies with author overlap compared to 64.6% success replication rates without author overlap).
+
Nobel laureate and professor emeritus in psychology Daniel Kahneman argued that the original authors should be involved in the replication effort because the published methods are often too vague.<ref name="Chambers">{{cite news|url=https://www.theguardian.com/science/head-quarters/2014/jun/10/physics-envy-do-hard-sciences-hold-the-solution-to-the-replication-crisis-in-psychology|title=Physics envy: Do 'hard' sciences hold the solution to the replication crisis in psychology?|first=Chris|last=Chambers|date=10 June 2014|newspaper=The Guardian}}</ref> Others such as Dr. Andrew Wilson disagree and argue that the methods should be written down in detail.<ref name="Chambers" /> An investigation of replication rates in psychology in 2012 indicated higher success rates of replication in replication studies when there was author overlap with the original authors of a study<ref name=":1">{{Cite journal|title = Replications in Psychology Research How Often Do They Really Occur?|journal = Perspectives on Psychological Science|date = 2012-11-01|issn = 1745-6916|pmid = 26168110|pages = 537–542|volume = 7|issue = 6|doi = 10.1177/1745691612460688|first = Matthew C.|last = Makel|first2 = Jonathan A.|last2 = Plucker|first3 = Boyd|last3 = Hegarty}}</ref> (91.7% successful replication rates in studies with author overlap compared to 64.6% success replication rates without author overlap).
  
====Psychology replication rates====
+
==== Psychology replication rates ====
A report by the [https://en.wikipedia.org/wiki/Center_for_Open_Science Open Science Collaboration] in August 2015 that was coordinated by [https://en.wikipedia.org/wiki/Brian_Nosek Brian Nosek] estimated the reproducibility of 100 studies in psychological science from three high-ranking psychology journals.<ref>{{Cite journal|title = Estimating the reproducibility of psychological science|journal = Science|date = 2015-08-28|issn = 0036-8075|pmid = 26315443|pages = aac4716|volume = 349|issue = 6251|doi = 10.1126/science.aac4716|first = Open Science|last = Collaboration|url = http://eprints.keele.ac.uk/877/1/Open%20Science%20%28Science%20Pre-Print%29.pdf|hdl = 10722/230596}}</ref> Overall, 36% of the replications yielded significant findings (''p'' value below 0.05) compared to 97% of the original studies that had significant effects. The mean [[effect size]] in the replications was approximately half the magnitude of the effects reported in the original studies.
 
  
The same paper examined the reproducibility rates and effect sizes by journal (''[[Journal of Personality and Social Psychology]]'' [JPSP], ''[[Journal of Experimental Psychology: Learning, Memory, and Cognition]]'' [JEP:LMC], ''[[Psychological Science (journal)|Psychological Science]]'' [PSCI]) and discipline ([[social psychology]], [[cognitive psychology]]). Study replication rates were 23% for JPSP, 38% for JEP:LMC, and 38% for PSCI. Studies in the field of cognitive psychology had a higher replication rate (50%) than studies in the field of social psychology (25%).
+
A report by the [https://en.wikipedia.org/wiki/Center_for_Open_Science Open Science Collaboration] in August 2015 that was coordinated by [https://en.wikipedia.org/wiki/Brian_Nosek Brian Nosek] estimated the reproducibility of 100 studies in psychological science from three high-ranking psychology journals.<ref>{{Cite journal|title = Estimating the reproducibility of psychological science|journal = Science|date = 2015-08-28|issn = 0036-8075|pmid = 26315443|pages = aac4716|volume = 349|issue = 6251|doi = 10.1126/science.aac4716|first = Open Science|last = Collaboration|url = http://eprints.keele.ac.uk/877/1/Open%20Science%20%28Science%20Pre-Print%29.pdf|hdl = 10722/230596}}</ref> Overall, 36% of the replications yielded significant findings (''p'' value below 0.05) compared to 97% of the original studies that had significant effects. The mean effect size in the replications was approximately half the magnitude of the effects reported in the original studies.
  
An analysis of the publication history in the top 100 psychology journals between 1900 and 2012 indicated that approximately 1.6% of all psychology publications were replication attempts.<ref name=":1" /> Articles were considered a replication attempt if the term "replication" appeared in the text. A subset of those studies (500 studies) was randomly selected for further examination and yielded a lower replication rate of 1.07% (342 of the 500 studies [68.4%]  were actually replications). In the subset of 500 studies, analysis indicated that 78.9% of published replication attempts were successful. The rate of successful replication was significantly higher when at least one author of the original study was part of the replication attempt (91.7% relative to 64.6%).
+
The same paper examined the reproducibility rates and effect sizes by journal (''Journal of Personality and Social Psychology'' [JPSP], ''Journal of Experimental Psychology: Learning, Memory, and Cognition'' [JEP:LMC], ''Psychological Science'' [PSCI]) and discipline (social psychology, cognitive psychology). Study replication rates were 23% for JPSP, 38% for JEP:LMC, and 38% for PSCI. Studies in the field of cognitive psychology had a higher replication rate (50%) than studies in the field of social psychology (25%).
  
A study published in 2018 in ''[[Nature (journal)|Nature Human Behaviour]]'' sought to replicate 21 social and behavioral science papers from ''Nature'' and ''[[Science (journal)|Science]],'' finding that only 13 could be successfully replicated.<ref>{{Cite news|url=https://www.wired.com/story/social-science-reproducibility|title=The Science Behind Social Science Gets Shaken Up—Again|work=WIRED|access-date=2018-08-28|language=en-US}}</ref><ref>{{cite journal |last1=Camerer |first1=Colin F. |authorlink1=Colin Camerer|last2=Dreber |first2=Anna |last3=Holzmeister |first3=Felix |last4=Ho |first4=Teck-Hua |last5=Huber |first5=Jürgen |last6=Johannesson |first6=Magnus |last7=Kirchler |first7=Michael |last8=Nave |first8=Gideon |last9=Nosek |first9=Brian A. |last10=Pfeiffer |first10=Thomas |last11=Altmejd |first11=Adam |last12=Buttrick |first12=Nick |last13=Chan |first13=Taizan |last14=Chen |first14=Yiling |last15=Forsell |first15=Eskil |last16=Gampa |first16=Anup |last17=Heikensten |first17=Emma |last18=Hummer |first18=Lily |last19=Imai |first19=Taisuke |last20=Isaksson |first20=Siri |last21=Manfredi |first21=Dylan |last22=Rose |first22=Julia |last23=Wagenmakers |first23=Eric-Jan |last24=Wu |first24=Hang |display-authors=2 |title=Evaluating the replicability of social science experiments in Nature and Science between 2010 and 2015 |journal=Nature Human Behaviour |date=27 August 2018 |volume=2 |issue=9 |pages=637–644 |doi=10.1038/s41562-018-0399-z|pmid=31346273 |url=https://authors.library.caltech.edu/91063/2/41562_2018_399_MOESM1_ESM.pdf }}</ref> Similarly, in a study conducted under the auspices of the [https://en.wikipedia.org/wiki/Center_for_Open_Science Center for Open Science], a team of 186 researchers from 60 different laboratories (representing 36 different nationalities from 6 different continents) conducted replications of 28 classic and contemporary findings in psychology.<ref name="Many Labs 2">{{cite journal |last1=Klein |first1=R.A. |title=Many Labs 2: Investigating Variation in Replicability Across Samples and Settings |journal=Advances in Methods and Practices in Psychological Science |date=2018 |volume=1 |issue=4 |pages=443–490 |doi=10.1177/2515245918810225 }}</ref> The focus of the study was not only on whether or not the findings from the original papers replicated, but also on the extent to which findings varied as a function of variations in samples and contexts. Overall, 14 of the 28 findings failed to replicate despite massive sample sizes. However, if a finding replicated, it replicated in most samples, while if a finding was not replicated, it failed to replicate with little variation across samples and contexts. This evidence is inconsistent with a popular explanation that failures to replicate in psychology are likely due to changes in the sample between the original and replication study.<ref name="Witkowski">{{cite journal |last1=Witkowski |first1=Tomasz |title=Is the glass half empty or half full? Latest results in the replication crisis in Psychology |journal=Skeptical Inquirer |date=2019 |volume=43 |issue=2 |pages=5–6}}</ref>
+
An analysis of the publication history in the top 100 psychology journals between 1900 and 2012 indicated that approximately 1.6% of all psychology publications were replication attempts.<ref name=":1" /> Articles were considered a replication attempt if the term "replication" appeared in the text. A subset of those studies (500 studies) was randomly selected for further examination and yielded a lower replication rate of 1.07% (342 of the 500 studies [68.4%] were actually replications). In the subset of 500 studies, analysis indicated that 78.9% of published replication attempts were successful. The rate of successful replication was significantly higher when at least one author of the original study was part of the replication attempt (91.7% relative to 64.6%).
 +
 
 +
A study published in 2018 in ''Nature Human Behaviour'' sought to replicate 21 social and behavioral science papers from ''Nature'' and ''Science,'' finding that only 13 could be successfully replicated.<ref>{{Cite news|url=https://www.wired.com/story/social-science-reproducibility|title=The Science Behind Social Science Gets Shaken Up—Again|work=WIRED|access-date=2018-08-28|language=en-US}}</ref><ref>{{cite journal |last1=Camerer |first1=Colin F. |authorlink1=Colin Camerer|last2=Dreber |first2=Anna |last3=Holzmeister |first3=Felix |last4=Ho |first4=Teck-Hua |last5=Huber |first5=Jürgen |last6=Johannesson |first6=Magnus |last7=Kirchler |first7=Michael |last8=Nave |first8=Gideon |last9=Nosek |first9=Brian A. |last10=Pfeiffer |first10=Thomas |last11=Altmejd |first11=Adam |last12=Buttrick |first12=Nick |last13=Chan |first13=Taizan |last14=Chen |first14=Yiling |last15=Forsell |first15=Eskil |last16=Gampa |first16=Anup |last17=Heikensten |first17=Emma |last18=Hummer |first18=Lily |last19=Imai |first19=Taisuke |last20=Isaksson |first20=Siri |last21=Manfredi |first21=Dylan |last22=Rose |first22=Julia |last23=Wagenmakers |first23=Eric-Jan |last24=Wu |first24=Hang |display-authors=2 |title=Evaluating the replicability of social science experiments in Nature and Science between 2010 and 2015 |journal=Nature Human Behaviour |date=27 August 2018 |volume=2 |issue=9 |pages=637–644 |doi=10.1038/s41562-018-0399-z|pmid=31346273 |url=https://authors.library.caltech.edu/91063/2/41562_2018_399_MOESM1_ESM.pdf }}</ref> Similarly, in a study conducted under the auspices of the [https://en.wikipedia.org/wiki/Center_for_Open_Science Center for Open Science], a team of 186 researchers from 60 different laboratories (representing 36 different nationalities from 6 different continents) conducted replications of 28 classic and contemporary findings in psychology.<ref name="Many Labs 2">{{cite journal |last1=Klein |first1=R.A. |title=Many Labs 2: Investigating Variation in Replicability Across Samples and Settings |journal=Advances in Methods and Practices in Psychological Science |date=2018 |volume=1 |issue=4 |pages=443–490 |doi=10.1177/2515245918810225 }}</ref> The focus of the study was not only on whether or not the findings from the original papers replicated, but also on the extent to which findings varied as a function of variations in samples and contexts. Overall, 14 of the 28 findings failed to replicate despite massive sample sizes. However, if a finding replicated, it replicated in most samples, while if a finding was not replicated, it failed to replicate with little variation across samples and contexts. This evidence is inconsistent with a popular explanation that failures to replicate in psychology are likely due to changes in the sample between the original and replication study.<ref name="Witkowski">{{cite journal |last1=Witkowski |first1=Tomasz |title=Is the glass half empty or half full? Latest results in the replication crisis in Psychology |journal=Skeptical Inquirer |date=2019 |volume=43 |issue=2 |pages=5–6}}</ref>
  
 
==== A disciplinary social dilemma ====
 
==== A disciplinary social dilemma ====
 +
 
Highlighting the social structure that discourages replication in psychology, Brian D. Earp and Jim A. C. Everett enumerated five points as to why replication attempts are uncommon:<ref>See also [http://journal.frontiersin.org/article/10.3389/fpsyg.2015.00621/abstract Earp and Trafimow], 2015</ref><ref name="Everett 1152">{{Cite journal|title = A tragedy of the (academic) commons: interpreting the replication crisis in psychology as a social dilemma for early-career researchers|journal = Frontiers in Psychology|date = 2015-01-01|pmc = 4527093|pmid = 26300832|pages = 1152|doi = 10.3389/fpsyg.2015.01152|first = Jim Albert Charlton|last = Everett|first2 = Brian D.|last2 = Earp|volume=6}}</ref>
 
Highlighting the social structure that discourages replication in psychology, Brian D. Earp and Jim A. C. Everett enumerated five points as to why replication attempts are uncommon:<ref>See also [http://journal.frontiersin.org/article/10.3389/fpsyg.2015.00621/abstract Earp and Trafimow], 2015</ref><ref name="Everett 1152">{{Cite journal|title = A tragedy of the (academic) commons: interpreting the replication crisis in psychology as a social dilemma for early-career researchers|journal = Frontiers in Psychology|date = 2015-01-01|pmc = 4527093|pmid = 26300832|pages = 1152|doi = 10.3389/fpsyg.2015.01152|first = Jim Albert Charlton|last = Everett|first2 = Brian D.|last2 = Earp|volume=6}}</ref>
# "Independent, direct replications of others' findings can be time-consuming for the replicating researcher
+
 
# "[Replications] are likely to take energy and resources directly away from other projects that reflect one's own original thinking
+
#"Independent, direct replications of others' findings can be time-consuming for the replicating researcher  
# "[Replications] are generally harder to publish (in large part because they are viewed as being unoriginal)
+
#"[Replications] are likely to take energy and resources directly away from other projects that reflect one's own original thinking  
# "Even if [replications] are published, they are likely to be seen as 'bricklaying' exercises, rather than as major contributions to the field
+
#"[Replications] are generally harder to publish (in large part because they are viewed as being unoriginal)  
# "[Replications] bring less recognition and reward, and even basic career security, to their authors"<ref name=":3">{{Cite web|title = Resolving the replication crisis in social psychology? A new proposal|url = http://www.spsp.org/blog/resolving-replication-crisis-social-psychology|publisher=Society for Personality and Social Psychology|accessdate = 2015-11-18}}</ref>  
+
#"Even if [replications] are published, they are likely to be seen as 'bricklaying' exercises, rather than as major contributions to the field  
 +
#"[Replications] bring less recognition and reward, and even basic career security, to their authors"<ref name=":3">{{Cite web|title = Resolving the replication crisis in social psychology? A new proposal|url = http://www.spsp.org/blog/resolving-replication-crisis-social-psychology|publisher=Society for Personality and Social Psychology|accessdate = 2015-11-18}}</ref>  
 +
 
 
For these reasons the authors advocated that psychology is facing a disciplinary social dilemma, where the interests of the discipline are at odds with the interests of the individual researcher.
 
For these reasons the authors advocated that psychology is facing a disciplinary social dilemma, where the interests of the discipline are at odds with the interests of the individual researcher.
  
==== "Methodological terrorism" controversy ====
+
=== Medicine ===
With the replication crisis of psychology earning attention, Princeton University psychologist Susan Fiske drew controversy for calling out critics of psychology.<ref name=":02">{{Cite news|url=http://uk.businessinsider.com/susan-fiske-methodological-terrorism-2016-9?r=US&IR=T|title=Scientists are furious after a famous psychologist accused her peers of 'methodological terrorism'|work=Business Insider|access-date=2017-10-04|language=en}}</ref><ref>{{Cite news|url=https://www.psychologicalscience.org/publications/observer/obsonline/draft-of-observer-column-sparks-strong-social-media-response.html|title=Draft of Observer Column Sparks Strong Social Media Response|work=Association for Psychological Science|access-date=2017-10-04|language=en-US}}</ref><ref>{{Cite journal|last=Fiske|first=Susan T.|date=2016-10-31|title=A Call to Change Science's Culture of Shaming|url=https://www.psychologicalscience.org/observer/a-call-to-change-sciences-culture-of-shaming|journal=APS Observer|language=en-US|volume=29|issue=9}}</ref><ref>{{Cite news|url=http://nymag.com/scienceofus/2016/10/inside-psychologys-methodological-terrorism-debate.html|title=Inside Psychology's 'Methodological Terrorism' Debate|last=Singal|first=Jesse|work=Science of Us|access-date=2017-10-04|language=en}}</ref> She called these unnamed "adversaries" names such as "methodological terrorist" and "self-appointed data police", and said that criticism of psychology should only be expressed in private or through contacting the journals.<ref name=":02" /> Columbia University statistician and political scientist [[Andrew Gelman]], "well-respected among the researchers driving the replication debate", responded to Fiske, saying that she had found herself willing to tolerate the "dead paradigm" of faulty statistics and had refused to retract publications even when errors were pointed out.<ref name=":02" /><ref>{{Cite news|url=http://andrewgelman.com/2017/10/04/breaking-pnas-changes-slogan/|title=BREAKING . . . . . . . PNAS updates its slogan! - Statistical Modeling, Causal Inference, and Social Science|date=2017-10-04|work=Statistical Modeling, Causal Inference, and Social Science|access-date=2017-10-04|language=en-US}}</ref> He added that her tenure as editor has been abysmal and that a number of published papers edited by her were found to be based on extremely weak statistics; one of Fiske's own published papers had a major statistical error and "impossible" conclusions.<ref name=":02" />
+
 
 +
Medical researchers were among the first ones to ring alarm regarding the lack of replicability of pre-clinical studies on drugs targeted for industrial use.<ref>[https://www.nature.com/articles/nrd3439-c1 Believe it or not: how much can we rely on published data on potential drug targets?] by Florian Prinz, Thomas Schlange & Khusru Asadullah, in: Nature Reviews Drug Discovery volume10, page712 (2011)</ref> This lack of reproducability and therefore of reliability in research started to become a growing problem for pharmaceutical companies because their decisions in which drug targets to invest money for expensive clinical research depends mainly on such studies.
 +
 
 +
Out of 49 medical studies from 1990–2003, with more than 1000 citations, 45 claimed that studied therapy was effective. Out of these studies, 16% were contradicted by subsequent studies, 16% had found stronger effects than did subsequent studies, 44% were replicated, and 24% remained largely unchallenged.<ref>{{cite journal|title=Contradicted and initially stronger effects in highly cited clinical research|last=Ioannidis JA|date=13 July 2005|journal=JAMA|volume=294|issue=2|pages=218–228|doi=10.1001/jama.294.2.218|pmid=16014596}}</ref> The US Food and Drug Administration in 1977–1990 found flaws in 10–20% of medical studies.<ref>{{cite journal|title=Scientific data audit—A key management tool|first=J. Leslie|last=Glick|date=1 January 1992|volume=2|issue=3|pages=153–168|doi=10.1080/08989629208573811|journal=Accountability in Research}}</ref> In a paper published in 2012, Glenn Begley, a biotech consultant working at Amgen, and Lee Ellis, at the University of Texas, argued that only 11% of the pre-clinical cancer studies could be replicated.<ref name="Begley-Lee2012">{{Cite journal|last1=Begley |first1=C. G. |last2=Ellis |first2=L. M. |year=2012 |title=Drug Development: Raise Standards for Preclinical Cancer Research |journal=Nature |volume=483 |issue=7391 |pages=531–533|doi=10.1038/483531a |pmid=22460880 |bibcode=2012Natur.483..531B }}</ref><ref name="Begley2013">{{Cite journal|last=Begley |first=C. G. |year=2013 |title=Reproducibility: Six red flags for suspect work |journal=Nature |volume=497 |issue=7450 |pages=433–434|doi=10.1038/497433a |bibcode=2013Natur.497..433B }}</ref>
  
=== In medicine ===
+
A 2016 article by [https://en.wikipedia.org/wiki/John_Ioannidis John Ioannidis], Professor of Medicine and of Health Research and Policy at Stanford University School of Medicine and a Professor of Statistics at Stanford University School of Humanities and Sciences, elaborated on "Why Most Clinical Research Is Not Useful".<ref name="Ioannidis2016">{{cite journal | last1 = Ioannidis | first1 = JPA | year = 2016 | title = Why Most Clinical Research Is Not Useful | journal = PLoS Med | volume = 13 | issue = 6| page = e1002049 | doi = 10.1371/journal.pmed.1002049 | pmid = 27328301 | pmc = 4915619 }}</ref> In the article Ioannidis laid out some of the problems and called for reform, characterizing certain points for medical research to be useful again; one example he made was the need for medicine to be "patient centered" (e.g. in the form of the Patient-Centered Outcomes Research Institute) instead of the current practice to mainly take care of "the needs of physicians, investigators, or sponsors". Ioannidis is known for his research focus on science itself since the 2005 paper "Why Most Published Research Findings Are False".<ref>{{cite journal|title=Why Most Published Research Findings Are False|journal=[[PLoS Medicine]]|date=August 1, 2005|issn=1549-1277|pmc=1182327|pmid=16060722|volume=2|issue=8|doi=10.1371/journal.pmed.0020124|first=John P. A.|last=Ioannidis|pages=e124}}</ref>
Out of 49 medical studies from 1990–2003, with more than 1000 citations, 45 claimed that studied therapy was effective. Out of these studies, 16% were contradicted by subsequent studies, 16% had found stronger effects than did subsequent studies, 44% were replicated, and 24% remained largely unchallenged.<ref>{{cite journal|title=Contradicted and initially stronger effects in highly cited clinical research|last=Ioannidis JA|date=13 July 2005|journal=JAMA|volume=294|issue=2|pages=218–228|doi=10.1001/jama.294.2.218|pmid=16014596}}</ref> The US [[Food and Drug Administration]] in 1977–1990 found flaws in 10–20% of medical studies.<ref>{{cite journal|title=Scientific data audit—A key management tool|first=J. Leslie|last=Glick|date=1 January 1992|volume=2|issue=3|pages=153–168|doi=10.1080/08989629208573811|journal=Accountability in Research}}</ref> In a paper published in 2012, Glenn Begley, a biotech consultant working at [[Amgen]], and Lee Ellis, at the University of Texas, argued that only 11% of the pre-clinical cancer studies could be replicated.<ref name="Begley-Lee2012">{{Cite journal|last1=Begley |first1=C. G. |last2=Ellis |first2=L. M. |year=2012 |title=Drug Development: Raise Standards for Preclinical Cancer Research |journal=Nature |volume=483 |issue=7391 |pages=531–533|doi=10.1038/483531a |pmid=22460880 |bibcode=2012Natur.483..531B }}</ref><ref name="Begley2013">{{Cite journal|last=Begley |first=C. G. |year=2013 |title=Reproducibility: Six red flags for suspect work |journal=Nature |volume=497 |issue=7450 |pages=433–434|doi=10.1038/497433a |bibcode=2013Natur.497..433B }}</ref>
 
  
A 2016 article by [[John Ioannidis]], Professor of Medicine and of Health Research and Policy at Stanford University School of Medicine and a Professor of Statistics at Stanford University School of Humanities and Sciences, elaborated on "Why Most Clinical Research Is Not Useful".<ref name="Ioannidis2016">{{cite journal | last1 = Ioannidis | first1 = JPA | year = 2016 | title = Why Most Clinical Research Is Not Useful | journal = PLoS Med | volume = 13 | issue = 6| page = e1002049 | doi = 10.1371/journal.pmed.1002049 | pmid = 27328301 | pmc = 4915619 }}</ref> In the article Ioannidis laid out some of the problems and called for reform, characterizing certain points for medical research to be useful again; one example he made was the need for medicine to be "patient centered" (e.g. in the form of the [[Patient-Centered Outcomes Research Institute]]) instead of the current practice to mainly take care of "the needs of physicians, investigators, or sponsors". Ioannidis is known for his research focus on science itself since the 2005 paper "Why Most Published Research Findings Are False".<ref>{{cite journal|title=Why Most Published Research Findings Are False|journal=[[PLoS Medicine]]|date=August 1, 2005|issn=1549-1277|pmc=1182327|pmid=16060722|volume=2|issue=8|doi=10.1371/journal.pmed.0020124|first=John P. A.|last=Ioannidis|pages=e124}}</ref>
+
=== Marketing ===
  
=== In marketing ===
+
Marketing is another discipline with a "desperate need" for replication.<ref>{{Cite journal|title = The desperate need for replications|journal = Journal of Consumer Research|date = 2001-06-01|pages = 149–158|volume = 28|issue = 1|doi = 10.1086/321953|first = John E.|last = Hunter}}</ref> Many famous marketing studies fail to be repeated upon replication, a notable example being the "too-many-choices" effect, in which a high number of choices of product makes a consumer less likely to purchase.<ref>{{Cite journal|last=Armstrong|first=J. S.|last2=Green|first2=Kesten|date=30 January 2017|title=Guidelines for Science: Evidence and Checklists|url=http://repository.upenn.edu/cgi/viewcontent.cgi?article=1258&context=marketing_papers|journal=Working Paper|volume=|pages=|via=}}</ref> In addition to the previously mentioned arguments, replications studies in marketing are needed to examine the applicability of theories and models across countries and cultures, which is especially important because of possible influences of globalization.<ref>{{Cite journal|title = Effects of Subcultural Differences on Country and Product Evaluations: A Replication Study|journal = Journal of Global Marketing|date = 2016-03-22|pages = 115–127|volume = 29|issue = 3|doi = 10.1080/08911762.2015.1138012|first1 = Thomas|last1 = Aichner|first2 = Paolo|last2 = Coletti|first3 = Cipriano|last3 = Forza|first4 = Urban|last4 = Perkmann|first5 = Alessio|last5 = Trentin}}</ref>
Marketing is another discipline with a "desperate need" for replication.<ref>{{Cite journal|title = The desperate need for replications|journal = Journal of Consumer Research|date = 2001-06-01|pages = 149–158|volume = 28|issue = 1|doi = 10.1086/321953|first = John E.|last = Hunter}}</ref> Many famous marketing studies fail to be repeated upon replication, a notable example being the "too-many-choices" effect, in which a high number of choices of product makes a consumer less likely to purchase.<ref>{{Cite journal|last=Armstrong|first=J. S.|last2=Green|first2=Kesten|date=30 January 2017|title=Guidelines for Science: Evidence and Checklists|url=http://repository.upenn.edu/cgi/viewcontent.cgi?article=1258&context=marketing_papers|journal=Working Paper|volume=|pages=|via=}}</ref> In addition to the previously mentioned arguments, replications studies in marketing are needed to examine the applicability of theories and models across countries and cultures, which is especially important because of possible influences of [[globalization]].<ref>{{Cite journal|title = Effects of Subcultural Differences on Country and Product Evaluations: A Replication Study|journal = Journal of Global Marketing|date = 2016-03-22|pages = 115–127|volume = 29|issue = 3|doi = 10.1080/08911762.2015.1138012|first1 = Thomas|last1 = Aichner|first2 = Paolo|last2 = Coletti|first3 = Cipriano|last3 = Forza|first4 = Urban|last4 = Perkmann|first5 = Alessio|last5 = Trentin}}</ref>
+
 
 +
=== Economics ===
  
=== In economics ===
 
 
A 2016 study in the journal ''Science'' found that one-third of 18 experimental studies from two top-tier economics journals (''American Economic Review'' and the ''Quarterly Journal of Economics'') failed to be successfully replicated.<ref>{{Cite journal|last=Camerer|first=Colin F.|last2=Dreber|first2=Anna|last3=Forsell|first3=Eskil|last4=Ho|first4=Teck-Hua|last5=Huber|first5=Jürgen|last6=Johannesson|first6=Magnus|last7=Kirchler|first7=Michael|last8=Almenberg|first8=Johan|last9=Altmejd|first9=Adam|date=2016-03-25|title=Evaluating replicability of laboratory experiments in economics|journal=Science|language=en|volume=351|issue=6280|pages=1433–1436|doi=10.1126/science.aaf0918|issn=0036-8075|pmid=26940865|bibcode=2016Sci...351.1433C}}</ref><ref>{{Cite news|url=http://www.sciencemag.org/news/2016/03/about-40-economics-experiments-fail-replication-survey|title=About 40% of economics experiments fail replication survey|date=2016-03-03|work=Science |access-date=2017-10-25|language=en}}</ref> A 2017 study in the ''Economic Journal'' suggested that "the majority of the average effects in the empirical economics literature are exaggerated by a factor of at least 2 and at least one-third are exaggerated by a factor of 4 or more".<ref>{{Cite journal|last=Ioannidis|first=John P. A.|last2=Stanley|first2=T. D.|last3=Doucouliagos|first3=Hristos|date=2017-10-01|title=The Power of Bias in Economics Research|journal=The Economic Journal|language=en|volume=127|issue=605|pages=F236–F265|doi=10.1111/ecoj.12461|issn=1468-0297}}</ref>
 
A 2016 study in the journal ''Science'' found that one-third of 18 experimental studies from two top-tier economics journals (''American Economic Review'' and the ''Quarterly Journal of Economics'') failed to be successfully replicated.<ref>{{Cite journal|last=Camerer|first=Colin F.|last2=Dreber|first2=Anna|last3=Forsell|first3=Eskil|last4=Ho|first4=Teck-Hua|last5=Huber|first5=Jürgen|last6=Johannesson|first6=Magnus|last7=Kirchler|first7=Michael|last8=Almenberg|first8=Johan|last9=Altmejd|first9=Adam|date=2016-03-25|title=Evaluating replicability of laboratory experiments in economics|journal=Science|language=en|volume=351|issue=6280|pages=1433–1436|doi=10.1126/science.aaf0918|issn=0036-8075|pmid=26940865|bibcode=2016Sci...351.1433C}}</ref><ref>{{Cite news|url=http://www.sciencemag.org/news/2016/03/about-40-economics-experiments-fail-replication-survey|title=About 40% of economics experiments fail replication survey|date=2016-03-03|work=Science |access-date=2017-10-25|language=en}}</ref> A 2017 study in the ''Economic Journal'' suggested that "the majority of the average effects in the empirical economics literature are exaggerated by a factor of at least 2 and at least one-third are exaggerated by a factor of 4 or more".<ref>{{Cite journal|last=Ioannidis|first=John P. A.|last2=Stanley|first2=T. D.|last3=Doucouliagos|first3=Hristos|date=2017-10-01|title=The Power of Bias in Economics Research|journal=The Economic Journal|language=en|volume=127|issue=605|pages=F236–F265|doi=10.1111/ecoj.12461|issn=1468-0297}}</ref>
  
=== In sports science ===
+
=== Sports Science ===
A 2018 study took the field of exercise and sports science to task for insufficient replication studies, limited reporting of null results and trivial results, and insufficient research transparency.<ref>{{Cite journal|last=Halperin|first=Israel|last2=Vigotsky|first2=Andrew D.|last3=Foster|first3=Carl|last4=Pyne|first4=David B.|date=2018-02-01|title=Strengthening the Practice of Exercise and Sport-Science Research|journal=International Journal of Sports Physiology and Performance|volume=13|issue=2|pages=127–134|doi=10.1123/ijspp.2017-0322|issn=1555-0273|pmid=28787228}}</ref> Statisticians have criticized sports science for common use of a controversial statistical method called "[[magnitude-based inference]]" that has allowed sports scientists to extract apparently significant results from noisy data where ordinary hypothesis testing would have found none.<ref>{{Cite news|url=https://fivethirtyeight.com/features/how-shoddy-statistics-found-a-home-in-sports-research/|title=How Shoddy Statistics Found A Home In Sports Research|date=2018-05-16|work=FiveThirtyEight|access-date=2018-05-16|language=en-US}}</ref>
 
  
=== In hydrology ===
+
A 2018 study took the field of exercise and sports science to task for insufficient replication studies, limited reporting of null results and trivial results, and insufficient research transparency.<ref>{{Cite journal|last=Halperin|first=Israel|last2=Vigotsky|first2=Andrew D.|last3=Foster|first3=Carl|last4=Pyne|first4=David B.|date=2018-02-01|title=Strengthening the Practice of Exercise and Sport-Science Research|journal=International Journal of Sports Physiology and Performance|volume=13|issue=2|pages=127–134|doi=10.1123/ijspp.2017-0322|issn=1555-0273|pmid=28787228}}</ref> Statisticians have criticized sports science for common use of a controversial statistical method called "magnitude-based inference" that has allowed sports scientists to extract apparently significant results from noisy data where ordinary hypothesis testing would have found none.<ref>{{Cite news|url=https://fivethirtyeight.com/features/how-shoddy-statistics-found-a-home-in-sports-research/|title=How Shoddy Statistics Found A Home In Sports Research|date=2018-05-16|work=FiveThirtyEight|access-date=2018-05-16|language=en-US}}</ref>
A 2019 study in ''Scientific Data'' found that a small number of articles in [[hydrology]] and [[water resources]] journals could be reproduced due to data unavailability. The study "estimated, with 95% confidence, that results might be reproduced for only 0.6% to 6.8% of all 1,989 articles".<ref>{{Cite journal|last=Stagge|first=James H.|last2=Rosenberg|first2=David E.|last3=Abdallah|first3=Adel M.|last4=Akbar|first4=Hadia|last5=Attallah|first5=Nour A.|last6=James|first6=Ryan|date=2019-02-26|title=Assessing data availability and research reproducibility in hydrology and water resources|journal=Scientific Data|language=en|volume=6|pages=190030|doi=10.1038/sdata.2019.30|pmid=30806638|pmc=6390703|issn=2052-4463|bibcode=2019NatSD...690030S}}</ref>
 
  
=== Political repercussions ===
+
=== Hydrology ===
In the US, science's reproducibility crisis has become a topic of political contention, linked to the attempt to diminish regulations – e.g. of emissions of pollutants, with the argument that these regulations are based on non-reproducible science.<ref name="Oreskes">{{Cite journal |last=Oreskes |first=N. |year=2018 |title=Beware: Transparency rule is a trojan horse |journal=Nature |volume=557 |issue=7706 |page=469 |doi=10.1038/d41586-018-05207-9|pmid=29789751 |bibcode=2018Natur.557..469O }}</ref><ref name="FUTURES2">{{cite journal |last1=Andrea |first1=Saltelli |title=Why science's crisis should not become a political battling ground |journal=Futures |date=December 2018 |volume=104 |pages=85–90 |doi=10.1016/j.futures.2018.07.006}}</ref> Previous attempts with the same aim accused studies of being non-transparent.<ref name="Michaels">{{Cite book|first=D. |last=Michaels |title=Doubt is their product: How industry's assault on science threatens your health |publisher=Oxford University Press |year=2008}}</ref>
+
 
 +
A 2019 study in ''Scientific Data'' found that a small number of articles in hydrology and water resources journals could be reproduced due to data unavailability. The study "estimated, with 95% confidence, that results might be reproduced for only 0.6% to 6.8% of all 1,989 articles".<ref>{{Cite journal|last=Stagge|first=James H.|last2=Rosenberg|first2=David E.|last3=Abdallah|first3=Adel M.|last4=Akbar|first4=Hadia|last5=Attallah|first5=Nour A.|last6=James|first6=Ryan|date=2019-02-26|title=Assessing data availability and research reproducibility in hydrology and water resources|journal=Scientific Data|language=en|volume=6|pages=190030|doi=10.1038/sdata.2019.30|pmid=30806638|pmc=6390703|issn=2052-4463|bibcode=2019NatSD...690030S}}</ref><ref>[https://www.nature.com/articles/sdata201930#f2 https://www.nature.com/articles/sdata201930#f2]</ref><ref>[https://replicationnetwork.com/2019/03/01/surveying-reproducibility/ https://replicationnetwork.com/2019/03/01/surveying-reproducibility/]</ref><br/> This study mainly relates to the quality of documentation of origninal material which is obviously only given scarcely or not at all. With no raw data available the factual validity of the scientific articles cannot be estimated, but the scientific non-reproducability is not proven thereby.
 +
 
 +
=== Hard Sciences ===
 +
 
 +
As opposed to the social and life sciences many agree that the “hard sciences” – physics, chemistry, biology, astronomy, etc. – will not be touched by this crisis. Nevertheless there seem to be quite a few problems there as well. „Even physics has been affected, as William Wilson notes. “Two of the most vaunted physics results of the past few years — the announced discovery of both cosmic inflation and gravitational waves at the BICEP2 experiment in Antarctica, and the supposed discovery of superluminal neutrinos at the Swiss-Italian border — have now been retracted, with far less fanfare than when they were first published.” See [http://phys.org/news/2015-02-cosmic-inflation-bicep2-results.html this about the former] and [http://press.cern/press-releases/2011/09/opera-experiment-reports-anomaly-flight-time-neutrinos-cern-gran-sasso this about the latter].”<ref>[ https://fabiusmaximus.com/2016/04/19/replication-crisis-in-science-95394/ The replication crisis in science has just begun. It will be big.] by Larry Kummer, Editor Science & Nature 19 April 2016 on Fabius Maximus website, here referring to W.Wilson – Scientific Regress (see Further readings)</ref>
 +
 
 +
=== Natural constants ===
 +
 
 +
One of the pillars of physics and all sciences based on the laws described in physics is the assumption of the constancy of the natural constants. This constancy is a general assumption because it can principally not been proven. One can only state that they have been constant since humans are able to measure them, which is an extremely short time span compared to the age of the universe.<br/> Nevertheless there is good reason to doubt even this basic assumption of the physical sciences.„At the end of 1998 the CODATA even decided to increase the uncertainty of the accepted value for the gravitational constant from 128 ppm to 1500 ppm. This remarkable step of increasing the uncertainty instead of decreasing was made to reflect the discrepancies between recent experiments, which span a wide range of more than 0.7 %.”<ref>Ulf Kleinevoß: Bestimmung der Newtonschen Gravitationskonstanten, Dissertation Januar 2002, Wuppertal, S.1, Abstract;
 +
[http://elpub.bib.uni-wuppertal.de/edocs/dokumente/fb08/diss2002/kleinevoss/d080201.pdf]</ref><br/> And in his work “The Science Delusion – Freeing the Spirit of Enquiry” biologist and philosopher [[Rupert_Sheldrake|Rupert Sheldrake]]<ref>Rupert Sheldrake: The Science Delusion – Freeing the Spirit of Enquiry; London 2012, Hodder& Stoughton, ISBN 978 1 444 72795 1. Chapter 3: Are the Laws of Nature Fixed?</ref> discusses this problem and gives lots of examples showing that the fundamental constants may not be as eternal as was thought but object to change by time. He prefers to see the “constants” more as habits of nature than as eternal laws. Sheldrake points out that the fundamental physical constants are artificially held as constant by defining them to be so and by reducing all measurements to the mean and eliminating strongly deviating measurings from the count. So as a matter of fact scientific measurings of the constants constantly produce quite different results, which are collected and regulated by the ''Committee on Data for Science and Technology'' ([https://howlingpixel.com/i-de/CODATA CODATA]) to always be constant as by definition.<ref>[https://howlingpixel.com/i-de/Physikalische_Konstante] and Sheldrake, Science Delusion</ref>
  
 
== Causes of the crisis ==
 
== Causes of the crisis ==
In a work published in 2015, Glenn Begley and [https://en.wikipedia.org/wiki/John_Ioannidis John Ioannidis] offer five bullets as to summarize the present predicaments:<ref name="Begley-Ioannidis2015">{{cite journal|author1=Begley, C. G.|author2=Ioannidis, J. P.|lastauthoramp=y|year=2015|title=Reproducibility in Science: Improving the Standard for Basic and Preclinical Research|journal=Circulation Research|volume=116|issue=1|pages=116–126|doi=10.1161/CIRCRESAHA.114.303819|pmid=25552691}}</ref>
 
  
* Generation of new data/publications at an unprecedented rate.
+
=== Bad Science ===
* Compelling evidence that the majority of these discoveries will not stand the test of time.
+
 
* Causes: failure to adhere to good scientific practice and the desperation to publish or perish.
+
The [https://en.wikipedia.org/wiki/Sokal_affair Sokal affair] and the discussion in its wake has shown clearly that there is a major problem of not distinguishing between real and fake science.<ref>Peter Boghossian, Ed.D. (aka Peter Boyle, Ed.D.), James Lindsay, Ph.D. (aka, Jamie Lindsay, Ph.D.): The Conceptual Penis as a Social Construct: A Sokal-Style Hoax on Gender Studies. SKEPTIC, 19.05.2017.[http://www.skeptic.com/reading_room/conceptual-penis-social-contruct-sokal-style-hoax-o]; Alexander Durin: Fehler im System mancher Wissenschaften. Telepolis, Heise, 02.03.2014. [https://www.heise.de/tp/features/Fehler-im-System-mancher-Wissenschaften-3502557.html]; Alan D. Sokal: Transgressing the Boundaries: Towards a Transformative Hermeneutics of Quantum Gravity. Social Text 46/47:217-252, 1996. [http://www.physics.nyu.edu/faculty/sokal/transgress_v2/transgress_v2_singlefile.html].</ref><br/> The editor of the prominent medical journal ''Lancet''<ref>Richard Horton – Offline: What is medicine’s 5 sigma? In: The Lancet VOLUME 385, ISSUE 9976, P1380, APRIL 11, 2015;
* This is a multifaceted, multistakeholder problem.
+
[https://www.thelancet.com/journals/lancet/article/PIIS0140-6736(15)60696-1/fulltext]</ref> Richard Horton writes about “apparent endemicity of bad research behaviour” and “bad scientific practices” when he points out that “much of the scientific literature, perhaps half, may simply be untrue.” As the editor of the probably most important medical journal he is in a position to recognize the problem.
* No single party is solely responsible, and no single solution will suffice.
+
 
+
=== High rate of publications ===
 +
 
 
In fact some predictions of a possible crisis in the quality control mechanism of science can be traced back several decades, especially among scholars in science and technology studies (STS). Derek de Solla Price – considered the father of scientometrics – predicted that science could reach 'senility' as a result of its own exponential growth.<ref name="DeSollaPrice1963">{{cite book|author1=De Solla Price|author2=Derek J.|year=1963|title=Little science big science|publisher=Columbia University Press}}</ref> Some present day literature seems to vindicate this 'overflow' prophesy, lamenting at decay in both attention and quality.<ref name="Siebert-al2015">{{cite journal|author1=Siebert, S.|author2=Machesky, L. M.|author3=Insall, R. H.|lastauthoramp=y|year=2015|title=Overflow in science and its implications for trust|journal=eLife|volume=4|page=e10825|doi=10.7554/eLife.10825|pmid=26365552|pmc=4563216}}</ref><ref name="Della-Briotta-al2015">{{cite journal|author1=Della Briotta Parolo, P.|author2=Kumar Pan|author3=R. Ghosh|author4=R. Huberman|author5=B.A. Kimmo Kaski|author6=Fortunato, S.|year=2015|title=Attention decay in science|journal=Journal of Informetrics|volume=9|issue=4|pages=734–745|doi=10.1016/j.joi.2015.07.006|arxiv=1503.01881|bibcode=2015arXiv150301881D}}</ref>
 
In fact some predictions of a possible crisis in the quality control mechanism of science can be traced back several decades, especially among scholars in science and technology studies (STS). Derek de Solla Price – considered the father of scientometrics – predicted that science could reach 'senility' as a result of its own exponential growth.<ref name="DeSollaPrice1963">{{cite book|author1=De Solla Price|author2=Derek J.|year=1963|title=Little science big science|publisher=Columbia University Press}}</ref> Some present day literature seems to vindicate this 'overflow' prophesy, lamenting at decay in both attention and quality.<ref name="Siebert-al2015">{{cite journal|author1=Siebert, S.|author2=Machesky, L. M.|author3=Insall, R. H.|lastauthoramp=y|year=2015|title=Overflow in science and its implications for trust|journal=eLife|volume=4|page=e10825|doi=10.7554/eLife.10825|pmid=26365552|pmc=4563216}}</ref><ref name="Della-Briotta-al2015">{{cite journal|author1=Della Briotta Parolo, P.|author2=Kumar Pan|author3=R. Ghosh|author4=R. Huberman|author5=B.A. Kimmo Kaski|author6=Fortunato, S.|year=2015|title=Attention decay in science|journal=Journal of Informetrics|volume=9|issue=4|pages=734–745|doi=10.1016/j.joi.2015.07.006|arxiv=1503.01881|bibcode=2015arXiv150301881D}}</ref>
  
Line 84: Line 102:
 
Historian Philip Mirowski offered more recently a similar diagnosis in his 2011 book ''Science Mart'' (2011).<ref name="Mirowski">{{cite book|author=Mirowski, P.|year=2011|title=Science-Mart: Privatizing American Science|publisher=Harvard University Press}}</ref> 'Mart' is here a reference to the retail giant 'Walmart' and an allusion to the commodification of science. In the analysis of Mirowski, when science becomes a commodity being traded in a market, its quality collapses. Mirowski argues his case by tracing the decay of science to the decision of major corporations to close their in-house laboratories in order to outsource their work to universities, and subsequently to move their research away from universities to even cheaper contract research organizations (CRO).
 
Historian Philip Mirowski offered more recently a similar diagnosis in his 2011 book ''Science Mart'' (2011).<ref name="Mirowski">{{cite book|author=Mirowski, P.|year=2011|title=Science-Mart: Privatizing American Science|publisher=Harvard University Press}}</ref> 'Mart' is here a reference to the retail giant 'Walmart' and an allusion to the commodification of science. In the analysis of Mirowski, when science becomes a commodity being traded in a market, its quality collapses. Mirowski argues his case by tracing the decay of science to the decision of major corporations to close their in-house laboratories in order to outsource their work to universities, and subsequently to move their research away from universities to even cheaper contract research organizations (CRO).
  
The crisis of science's quality control system is affecting the use of science for policy. This is the thesis of a recent work by a group of STS scholars, who identify in 'evidence based (or informed) policy' a point of present tension.<ref name="FUTURES1">{{Cite journal|last1=Saltelli |first1=A. |last2=Funtowicz |first2=S. |year=2017 |title=What is science's crisis really about? |journal=Futures |volume=91 |pages=5–11 |doi=10.1016/j.futures.2017.05.010}}</ref><ref name="Benessia-al2016">{{Cite book|last1=Benessia |first1=A. |last2=Funtowicz |first2=S. |last3=Giampietro |first3=M. |last4=Guimarães Pereira |first4=A. |last5=Ravetz |first5=J. |last6=Saltelli |first6=A. |last7=Strand |first7=R. |last8=van der Sluijs |first8=J. |year=2016 |title=The Rightful Place of Science: Science on the Verge |publisher=Consortium for Science, Policy and Outcomes at Arizona State University|title-link=The Rightful Place of Science: Science on the Verge }}</ref><ref name="Letter">{{cite magazine|first1=Andrea |last1=Saltelli|first2=Jerome R. |last2=Ravetz|first3=Silvio |last3=Funtowicz|lastauthoramp=y|url=https://www.newscientist.com/letter/mg23030791-600-7-a-new-community-for-science/|title=A new community for science|magazine=New Scientist|issue=3079|p=52|date=25 June 2016}}</ref><ref name="FUTURES2" /> Economist Noah Smith suggests that a factor in the crisis has been the overvaluing of research in academia and undervaluing of teaching ability, especially in fields with few major recent discoveries.<ref name="Academic signaling and the post-truth world">{{cite web|last1=Smith|first1=Noah|title=Academic signaling and the post-truth world|url=http://noahpinionblog.blogspot.co.uk/2016/12/academic-signaling-and-post-truth-world.html|website=Noahpinion|publisher=Stony Brook University|accessdate=5 November 2017|date=2016-12-14}}</ref>
+
=== Insufficient control ===
  
Another major cause is [[publication bias]], the fact that positive results are more likely to get published than negative (or null) results. This may lead to the canonization of false facts <ref>{{cite journal |last1 = Nissen |first1 = Silas Boye |last2 = Magidson| first2 = Tali| last3 = Gross| first3 = Kevin | last4 = Bergstrom| first4 = Carl |title=Research: Publication bias and the canonization of false facts |journal=eLife |date=December 20, 2016 |volume=5 |pages=e21451 |url= https://elifesciences.org/articles/21451 |accessdate=9 June 2019 |doi=10.7554/eLife.21451 |arxiv= 1609.00494 }}</ref>.
+
The crisis of science's quality control system is affecting the use of science for policy. This is the thesis of a recent work by a group of STS scholars, who identify in 'evidence based (or informed) policy' a point of present tension.<ref name="FUTURES1">{{Cite journal|last1=Saltelli |first1=A. |last2=Funtowicz |first2=S. |year=2017 |title=What is science's crisis really about? |journal=Futures |volume=91 |pages=5–11 |doi=10.1016/j.futures.2017.05.010}}</ref><ref name="Benessia-al2016">{{Cite book|last1=Benessia |first1=A. |last2=Funtowicz |first2=S. |last3=Giampietro |first3=M. |last4=Guimarães Pereira |first4=A. |last5=Ravetz |first5=J. |last6=Saltelli |first6=A. |last7=Strand |first7=R. |last8=van der Sluijs |first8=J. |year=2016 |title=The Rightful Place of Science: Science on the Verge |publisher=Consortium for Science, Policy and Outcomes at Arizona State University|title-link=The Rightful Place of Science: Science on the Verge }}</ref><ref name="Letter">{{cite magazine|first1=Andrea |last1=Saltelli|first2=Jerome R. |last2=Ravetz|first3=Silvio |last3=Funtowicz|lastauthoramp=y|url=https://www.newscientist.com/letter/mg23030791-600-7-a-new-community-for-science/|title=A new community for science|magazine=New Scientist|issue=3079|p=52|date=25 June 2016}}</ref><ref name="FUTURES2" /> Economist Noah Smith suggests that a factor in the crisis has been the overvaluing of research in academia and undervaluing of teaching ability, especially in fields with few major recent discoveries.<ref name="Academic signaling and the post-truth world">{{cite web|last1=Smith|first1=Noah|title=Academic signaling and the post-truth world|url=http://noahpinionblog.blogspot.co.uk/2016/12/academic-signaling-and-post-truth-world.html|website=Noahpinion|publisher=Stony Brook University|accessdate=5 November 2017|date=2016-12-14}}</ref> In the journal ''Science'' article “Who’s Afraid of Peer Review?”<ref> [https://science.sciencemag.org/content/342/6154/60.full John Bohannon: Who’s Afraid of Peer Review?], Science  04 Oct 2013: Vol. 342, Issue 6154, pp. 60-65;  DOI: 10.1126/science.342.6154.60</ref> author John Bohannon explains the problem of the growing number of open access journals and there lack of effective peer review. Of 300 fake articles filled with obvious mistakes and nonsense sent out to as many different online journals more than half were accepted without relevant criticism.
 +
 
 +
=== Publication bias ===
 +
 
 +
Another major cause is publication bias, the fact that positive results are more likely to get published than negative (or null) results. This may lead to the canonization of false facts <ref>{{cite journal |last1 = Nissen |first1 = Silas Boye |last2 = Magidson| first2 = Tali| last3 = Gross| first3 = Kevin | last4 = Bergstrom| first4 = Carl |title=Research: Publication bias and the canonization of false facts |journal=eLife |date=December 20, 2016 |volume=5 |pages=e21451 |url= https://elifesciences.org/articles/21451 |accessdate=9 June 2019 |doi=10.7554/eLife.21451 |arxiv= 1609.00494 }}</ref>.
 +
 
 +
=== Fundamental causes in the scientific paradigm ===
 +
 
 +
All the above reasons for lack of replicability have one thing in common: They are consequences of human fault, misconduct or mistakes and could theoretically be corrected by discipline, correctness and better control. Nevertheless there remains the possibility that – apart from all those considerations and observation being correct – there could also be more fundamental problems of the axiom of replicability and the constancy of nature as such.<br/> The biologist, philosopher and meta-scientist [[Rupert_Sheldrake|Rupert Sheldrake]] has pointed out that there is ample reason to doubt the belief in the full replicability and reliability of the hard sciences as well and that there might be a problem with the assumption of replicability as such.<ref>Rupert Sheldrake: [https://www.sheldrake.org/essays/how-the-universal-gravitational-constant-varies?highlight=WyJwaHlzaWNhbCIsImNvbnN0YW50cyIsInBoeXNpY2FsIGNvbnN0YW50cyJd How the Universal Gravitational Constant Varies];
 +
[https://www.sheldrake.org/essays/the-replicability-crisis-in-science The Replicability Crisis in Science]; 
 +
Science Set Free: 10 Paths to New Discovery, ISBN-13: 978-0770436704, Publ. Deepak Chopra; 1st ed. (4. September 2012); The Science Delusion, ISBN-13: 978-1444727944, Publisher: Coronet (6 Dec. 2012) </ref>
 +
 
 +
In any case, the [[Decline_Effect|Decline effect (see main article)]] indicates general problems of replicability because it cannot be fully explained by the causes proposed so far. It was discovered by J.B.Rhine in the 20s of the last century, long before the current crisis and under completely different circumstances. Rhine found the effect in his own research, even in areas whose statistical validity he himself had previously verified.
 +
 
 +
== Public Response ==
 +
 
 +
=== Political Repercussions ===
 +
 
 +
In the US, science's reproducibility crisis has become a topic of political contention, linked to the attempt to diminish regulations – e.g. of emissions of pollutants, with the argument that these regulations are based on non-reproducible science.<ref name="Oreskes">{{Cite journal |last=Oreskes |first=N. |year=2018 |title=Beware: Transparency rule is a trojan horse |journal=Nature |volume=557 |issue=7706 |page=469 |doi=10.1038/d41586-018-05207-9|pmid=29789751 |bibcode=2018Natur.557..469O }}</ref><ref name="FUTURES2">{{cite journal |last1=Andrea |first1=Saltelli |title=Why science's crisis should not become a political battling ground |journal=Futures |date=December 2018 |volume=104 |pages=85–90 |doi=10.1016/j.futures.2018.07.006}}</ref> Previous attempts with the same aim accused studies of being non-transparent.<ref name="Michaels">{{Cite book|first=D. |last=Michaels |title=Doubt is their product: How industry's assault on science threatens your health |publisher=Oxford University Press |year=2008}}</ref>
 +
 
 +
=== Public discussion ===
 +
 
 +
Apart from studies and articles in scientific journals there has been little public discussion of the Replicability Crisis as yet and few reports in the mass media. Most of the discussions are related to statistical, economical and sociological reasons only.<ref>See all of the cited literature.</ref><br/> The question whether there is a general problem of replicability behind these findings, too, has not been tackled so far. And there has been very little effort in the hard sciences to deal with the replication issue as there seems to be a consense that the problem is only one of the social and life sciences.
  
 
== Addressing the replication crisis ==
 
== Addressing the replication crisis ==
[[Reproducibility|Replication]] has been referred to as "the cornerstone of science".<ref>{{Cite journal|title = Most Published Research Findings Are False—But a Little Replication Goes a Long Way|journal = PLoS Med|date = 2007-02-27|pmc = 1808082|pmid = 17326704|pages = e28|volume = 4|issue = 2|doi = 10.1371/journal.pmed.0040028|first = Ramal|last = Moonesinghe|first2 = Muin J.|last2 = Khoury|first3 = A. Cecile J. W.|last3 = Janssens}}</ref><ref>{{Cite journal|title = The Value of Direct Replication|journal = Perspectives on Psychological Science|date = 2014-01-01|issn = 1745-6916|pmid = 26173243|pages = 76–80|volume = 9|issue = 1|doi = 10.1177/1745691613514755|first = Daniel J.|last = Simons}}</ref> Replication studies attempt to evaluate whether published results reflect true findings or false positives. The integrity of scientific findings and reproducibility of research are important as they form the knowledge foundation on which future studies are built.
 
  
===Metascience===
+
Replication has been referred to as "the cornerstone of science".<ref>{{Cite journal|title = Most Published Research Findings Are False—But a Little Replication Goes a Long Way|journal = PLoS Med|date = 2007-02-27|pmc = 1808082|pmid = 17326704|pages = e28|volume = 4|issue = 2|doi = 10.1371/journal.pmed.0040028|first = Ramal|last = Moonesinghe|first2 = Muin J.|last2 = Khoury|first3 = A. Cecile J. W.|last3 = Janssens}}</ref><ref>{{Cite journal|title = The Value of Direct Replication|journal = Perspectives on Psychological Science|date = 2014-01-01|issn = 1745-6916|pmid = 26173243|pages = 76–80|volume = 9|issue = 1|doi = 10.1177/1745691613514755|first = Daniel J.|last = Simons}}</ref> Replication studies attempt to evaluate whether published results reflect true findings or false positives. The integrity of scientific findings and reproducibility of research are important as they form the knowledge foundation on which future studies are built.
Metascience is the use of scientific methodology to study science itself. Metascience seeks to increase the quality of scientific research while reducing waste. It is also known as "''research on research''" and "''the science of science''", as it uses research methods to study how research is done and where improvements can be made. Metascience concerns itself with all fields of research and has been described as "a bird's eye view of science."<ref>{{Cite journal| doi = 10.1371/journal.pbio.1002264| pmid = 26431313| pmc = 4592065| issn = 1545-7885| volume = 13| issue = 10| pages = –1002264| last1 = Ioannidis| first1 = John P. A.| last2 = Fanelli| first2 = Daniele| last3 = Dunne| first3 = Debbie Drake| last4 = Goodman| first4 = Steven N.| title = Meta-research: Evaluation and Improvement of Research Methods and Practices| journal = PLOS Biology| date = 2015-10-02}}</ref> In the words of [[John Ioannidis]], "Science is the best thing that has happened to human beings ... but we can do it better."<ref>{{cite web |last1=Bach |first1=Author Becky |title=On communicating science and uncertainty: A podcast with John Ioannidis |url=https://scopeblog.stanford.edu/2015/12/08/on-communicating-science-and-uncertainty-a-podcast-with-john-ioannidis/ |website=Scope |accessdate=20 May 2019 |date=8 December 2015}}</ref>
 
  
Meta-research continues to be conducted to identify the roots of the crisis and to address them. Methods of addressing the crisis include [[pre-registration (science)|pre-registration]] of scientific studies and [[Clinical trial registration|clinical trials]] as well as the founding of organizations such as [[CONSORT]] and the [[EQUATOR Network]] that issue guidelines for methodology and reporting. There are continuing efforts to reform the system of academic incentives, to improve the [[academic peer review|peer review]] process, to reduce the [[misuse of statistics]], to combat [[bias]] in scientific literature, and to increase the overall quality and efficiency of the scientific process.
+
=== Metascience ===
 +
 
 +
Metascience is the use of scientific methodology to study science itself. Metascience seeks to increase the quality of scientific research while reducing waste. It is also known as "''research on research''" and "''the science of science''", as it uses research methods to study how research is done and where improvements can be made. Metascience concerns itself with all fields of research and has been described as "a bird's eye view of science."<ref>{{Cite journal| doi = 10.1371/journal.pbio.1002264| pmid = 26431313| pmc = 4592065| issn = 1545-7885| volume = 13| issue = 10| pages = –1002264| last1 = Ioannidis| first1 = John P. A.| last2 = Fanelli| first2 = Daniele| last3 = Dunne| first3 = Debbie Drake| last4 = Goodman| first4 = Steven N.| title = Meta-research: Evaluation and Improvement of Research Methods and Practices| journal = PLOS Biology| date = 2015-10-02}}</ref> In the words of John Ioannidis, "Science is the best thing that has happened to human beings ... but we can do it better."<ref>{{cite web |last1=Bach |first1=Author Becky |title=On communicating science and uncertainty: A podcast with John Ioannidis |url=https://scopeblog.stanford.edu/2015/12/08/on-communicating-science-and-uncertainty-a-podcast-with-john-ioannidis/ |website=Scope |accessdate=20 May 2019 |date=8 December 2015}}</ref>
 +
 
 +
Meta-research continues to be conducted to identify the roots of the crisis and to address them. Methods of addressing the crisis include pre-registration of scientific studies and clinical trials as well as the founding of organizations such as CONSORT and the EQUATOR Network that issue guidelines for methodology and reporting. There are continuing efforts to reform the system of academic incentives, to improve the peer review process, to reduce the misuse of statistics, to combat bias in scientific literature, and to increase the overall quality and efficiency of the scientific process.
  
 
=== Tackling publication bias with pre-registration of studies ===
 
=== Tackling publication bias with pre-registration of studies ===
A recent innovation in scientific publishing to address the replication crisis is through the use of [[registered report]]s.<ref>{{Cite web|title = Registered Replication Reports|publisher=Association for Psychological Science|url = http://www.psychologicalscience.org/index.php/replication|accessdate = 2015-11-13}}</ref><ref>{{Cite news|title = Psychology's 'registration revolution'|url = https://www.theguardian.com/science/head-quarters/2014/may/20/psychology-registration-revolution|newspaper = The Guardian|accessdate = 2015-11-13|first = Chris|last = Chambers|date = 2014-05-20}}</ref> The registered report format requires authors to submit a description of the study methods and analyses prior to data collection. Once the method and analysis plan is vetted through peer-review, publication of the findings is provisionally guaranteed, based on whether the authors follow the proposed protocol. One goal of registered reports is to circumvent the [[publication bias]] toward significant findings that can lead to implementation of questionable research practices and to encourage publication of studies with rigorous methods.
 
  
The journal ''[[Psychological Science (journal)|Psychological Science]]'' has encouraged the preregistration of studies and the reporting of effect sizes and confidence intervals.<ref>{{Cite journal|title = Replication in Psychological Science|journal = Psychological Science|date = 2015-11-09|issn = 0956-7976|pmid = 26553013|pages = 1827–32|doi = 10.1177/0956797615616374|first = D. Stephen|last = Lindsay|volume = 26|issue = 12}}</ref> The editor in chief also noted that the editorial staff will be asking for replication of studies with surprising findings from examinations using small sample sizes before allowing the manuscripts to be published.
+
A recent innovation in scientific publishing to address the replication crisis is through the use of registered reports.<ref>{{Cite web|title = Registered Replication Reports|publisher=Association for Psychological Science|url = http://www.psychologicalscience.org/index.php/replication|accessdate = 2015-11-13}}</ref><ref>{{Cite news|title = Psychology's 'registration revolution'|url = https://www.theguardian.com/science/head-quarters/2014/may/20/psychology-registration-revolution|newspaper = The Guardian|accessdate = 2015-11-13|first = Chris|last = Chambers|date = 2014-05-20}}</ref> The registered report format requires authors to submit a description of the study methods and analyses prior to data collection. Once the method and analysis plan is vetted through peer-review, publication of the findings is provisionally guaranteed, based on whether the authors follow the proposed protocol. One goal of registered reports is to circumvent the publication bias toward significant findings that can lead to implementation of questionable research practices and to encourage publication of studies with rigorous methods.
 +
 
 +
The journal ''Psychological Science'' has encouraged the preregistration of studies and the reporting of effect sizes and confidence intervals.<ref>{{Cite journal|title = Replication in Psychological Science|journal = Psychological Science|date = 2015-11-09|issn = 0956-7976|pmid = 26553013|pages = 1827–32|doi = 10.1177/0956797615616374|first = D. Stephen|last = Lindsay|volume = 26|issue = 12}}</ref> The editor in chief also noted that the editorial staff will be asking for replication of studies with surprising findings from examinations using small sample sizes before allowing the manuscripts to be published.
  
 
Moreover, only a very small proportion of academic journals in psychology and neurosciences explicitly stated that they welcome submissions of replication studies in their aim and scope or instructions to authors.<ref>{{Cite journal|last=Yeung|first=Andy W. K.|date=2017|title=Do Neuroscience Journals Accept Replications? A Survey of Literature|journal=Frontiers in Human Neuroscience|language=English|volume=11|pages=468|doi=10.3389/fnhum.2017.00468|pmid=28979201|pmc=5611708|issn=1662-5161}}</ref><ref>{{Cite journal|last=Martin|first=G. N.|last2=Clarke|first2=Richard M.|date=2017|title=Are Psychology Journals Anti-replication? A Snapshot of Editorial Practices|journal=Frontiers in Psychology|language=English|volume=8|pages=523|doi=10.3389/fpsyg.2017.00523|pmid=28443044|pmc=5387793|issn=1664-1078}}</ref> This phenomenon does not encourage the reporting or even attempt on replication studies.
 
Moreover, only a very small proportion of academic journals in psychology and neurosciences explicitly stated that they welcome submissions of replication studies in their aim and scope or instructions to authors.<ref>{{Cite journal|last=Yeung|first=Andy W. K.|date=2017|title=Do Neuroscience Journals Accept Replications? A Survey of Literature|journal=Frontiers in Human Neuroscience|language=English|volume=11|pages=468|doi=10.3389/fnhum.2017.00468|pmid=28979201|pmc=5611708|issn=1662-5161}}</ref><ref>{{Cite journal|last=Martin|first=G. N.|last2=Clarke|first2=Richard M.|date=2017|title=Are Psychology Journals Anti-replication? A Snapshot of Editorial Practices|journal=Frontiers in Psychology|language=English|volume=8|pages=523|doi=10.3389/fpsyg.2017.00523|pmid=28443044|pmc=5387793|issn=1664-1078}}</ref> This phenomenon does not encourage the reporting or even attempt on replication studies.
  
 
=== Emphasizing replication attempts in teaching ===
 
=== Emphasizing replication attempts in teaching ===
Based on coursework in experimental methods at MIT and Stanford, it has been suggested that methods courses in psychology emphasize replication attempts rather than original studies.<ref>{{Cite journal|title = Teaching Replication|journal = Perspectives on Psychological Science|date = 2012-11-01|issn = 1745-6916|pmid = 26168118|pages = 600–604|volume = 7|issue = 6|doi = 10.1177/1745691612460686|first = Michael C.|last = Frank|first2 = Rebecca|last2 = Saxe}}</ref><ref>{{Cite journal|title = Harnessing the Undiscovered Resource of Student Research Projects|journal = Perspectives on Psychological Science|date = 2012-11-01|issn = 1745-6916|pmid = 26168119|pages = 605–607|volume = 7|issue = 6|doi = 10.1177/1745691612459057|first = Jon E.|last = Grahe|first2 = Alan|last2 = Reifman|first3 = Anthony D.|last3 = Hermann|first4 = Marie|last4 = Walker|first5 = Kathryn C.|last5 = Oleson|first6 = Michelle|last6 = Nario-Redmond|first7 = Richard P.|last7 = Wiebe}}</ref> Such an approach would help students learn scientific methodology and provide numerous independent replications of meaningful scientific findings that would test the replicability of scientific findings. Some have recommended that graduate students should be required to publish a high-quality replication attempt on a topic related to their doctoral research prior to graduation.<ref name="Everett 1152"/>
+
 
 +
Based on coursework in experimental methods at MIT and Stanford, it has been suggested that methods courses in psychology emphasize replication attempts rather than original studies.<ref>{{Cite journal|title = Teaching Replication|journal = Perspectives on Psychological Science|date = 2012-11-01|issn = 1745-6916|pmid = 26168118|pages = 600–604|volume = 7|issue = 6|doi = 10.1177/1745691612460686|first = Michael C.|last = Frank|first2 = Rebecca|last2 = Saxe}}</ref><ref>{{Cite journal|title = Harnessing the Undiscovered Resource of Student Research Projects|journal = Perspectives on Psychological Science|date = 2012-11-01|issn = 1745-6916|pmid = 26168119|pages = 605–607|volume = 7|issue = 6|doi = 10.1177/1745691612459057|first = Jon E.|last = Grahe|first2 = Alan|last2 = Reifman|first3 = Anthony D.|last3 = Hermann|first4 = Marie|last4 = Walker|first5 = Kathryn C.|last5 = Oleson|first6 = Michelle|last6 = Nario-Redmond|first7 = Richard P.|last7 = Wiebe}}</ref> Such an approach would help students learn scientific methodology and provide numerous independent replications of meaningful scientific findings that would test the replicability of scientific findings. Some have recommended that graduate students should be required to publish a high-quality replication attempt on a topic related to their doctoral research prior to graduation.<ref name="Everett 1152" />
  
 
=== Reducing the ''p''-value required for claiming significance of new results ===
 
=== Reducing the ''p''-value required for claiming significance of new results ===
  
Many publications require a ''p''-value of ''p'' < 0.05 to claim statistical significance. The paper "Redefine statistical significance",<ref>{{Cite journal|title = Redefine statistical significance|url = https://psyarxiv.com/mky9j|journal = PsyArXiv|first = Daniel|last = Benjamin}}</ref> signed by a large number of scientists and mathematicians, proposes that in "fields where the threshold for defining statistical significance for new discoveries is ''P'' < 0.05, we propose a change to ''P'' < 0.005. This simple step would immediately improve the reproducibility of scientific research in many fields."
+
Many publications require a ''p''-value of ''p'' < 0.05 to claim statistical significance. The paper "Redefine statistical significance",<ref>{{Cite journal|title = Redefine statistical significance|url = https://psyarxiv.com/mky9j|journal = PsyArXiv|first = Daniel|last = Benjamin}}</ref> signed by a large number of scientists and mathematicians, proposes that in "fields where the threshold for defining statistical significance for new discoveries is ''P'' < 0.05, we propose a change to ''P'' < 0.005. This simple step would immediately improve the reproducibility of scientific research in many fields."
  
 
Their rationale is that "a leading cause of non-reproducibility (is that the) statistical standards of evidence for claiming new discoveries in many fields of science are simply too low. Associating 'statistically significant' findings with ''P'' < 0.05 results in a high rate of false positives even in the absence of other experimental, procedural and reporting problems."
 
Their rationale is that "a leading cause of non-reproducibility (is that the) statistical standards of evidence for claiming new discoveries in many fields of science are simply too low. Associating 'statistically significant' findings with ''P'' < 0.05 results in a high rate of false positives even in the absence of other experimental, procedural and reporting problems."
  
 
=== Addressing the misinterpretation of ''p''-values ===
 
=== Addressing the misinterpretation of ''p''-values ===
Although statisticians are unanimous that use of the ''p'' < 0.05 provides weaker evidence than is generally appreciated, there is a lack of unanimity about what should be done about it.  Some have advocated that Bayesian methods should replace ''p''-values. This has not happened on a wide scale, partly because it is complicated, and partly because many users distrust the specification of prior distributions in the absence of hard data. A simplified version of the Bayesian argument, based on testing a point null hypothesis was suggested by Colquhoun (2014, 2017).<ref name=DC2014>{{cite journal|last1=Colquhoun|first1=David|title=An investigation of the false discovery rate and the misinterpretation of ''p''-values|journal=Royal Society Open Science|year=2015|volume=1|issue=3|page=140216|doi=10.1098/rsos.140216|pmid=26064558|pmc=4448847|arxiv=1407.5296|bibcode=2014RSOS....140216C}}</ref><ref name="DC2017">{{cite journal|last1=Colquhoun|first1=David|title=The reproducibility of research and the misinterpretation of p-values|journal=Royal Society Open Science|volume=4|issue=12|pages=171085|date=2017|doi=10.1098/rsos.171085|pmid=29308247|pmc=5750014}}</ref> The logical problems of inductive inference were discussed in "The problem with p-values" (2016).<ref name="DC2016">{{cite web|last1=Colquhoun|first1=David|title=The problem with p-values|url=https://aeon.co/essays/it-s-time-for-science-to-abandon-the-term-statistically-significant|work=Aeon Magazine|accessdate=11 December 2016}}</ref>
 
  
The hazards of reliance on ''p''-values were emphasized by pointing out that even observation of ''p'' = 0.001 was not necessarily strong evidence against the null hypothesis.<ref name="DC2017"/> Despite the fact that the likelihood ratio in favour of the alternative hypothesis over the null is close to 100, if the hypothesis was implausible, with a prior probability of a real effect being 0.1, even the observation of ''p''&nbsp;=&nbsp;0.001 would have a false positive risk of 8 percent. It would not even reach the 5 percent level.
+
Although statisticians are unanimous that use of the ''p'' < 0.05 provides weaker evidence than is generally appreciated, there is a lack of unanimity about what should be done about it. Some have advocated that Bayesian methods should replace ''p''-values. This has not happened on a wide scale, partly because it is complicated, and partly because many users distrust the specification of prior distributions in the absence of hard data. A simplified version of the Bayesian argument, based on testing a point null hypothesis was suggested by Colquhoun (2014, 2017).<ref name="DC2014">{{cite journal|last1=Colquhoun|first1=David|title=An investigation of the false discovery rate and the misinterpretation of ''p''-values|journal=Royal Society Open Science|year=2015|volume=1|issue=3|page=140216|doi=10.1098/rsos.140216|pmid=26064558|pmc=4448847|arxiv=1407.5296|bibcode=2014RSOS....140216C}}</ref><ref name="DC2017">{{cite journal|last1=Colquhoun|first1=David|title=The reproducibility of research and the misinterpretation of p-values|journal=Royal Society Open Science|volume=4|issue=12|pages=171085|date=2017|doi=10.1098/rsos.171085|pmid=29308247|pmc=5750014}}</ref> The logical problems of inductive inference were discussed in "The problem with p-values" (2016).<ref name="DC2016">{{cite web|last1=Colquhoun|first1=David|title=The problem with p-values|url=https://aeon.co/essays/it-s-time-for-science-to-abandon-the-term-statistically-significant|work=Aeon Magazine|accessdate=11 December 2016}}</ref>
  
It was recommended<ref name="DC2017"/> that the terms "significant" and "non-significant" should not be used. ''p''-values and confidence intervals should still be specified, but they should be accompanied by an indication of the false positive risk. It was suggested that the best way to do this is to calculate the prior probability that would be necessary to believe in order to achieve a false positive risk of, say, 5%. The calculations can be done with [[R (programming language)|R]] scripts that are provided,<ref name="DC2017"/> or, more simply, with a web calculator.<ref name="fpr-calc">{{cite web|last1=Longstaff|first1=Colin|last2=Colquhoun|first2=David|title=Calculator for false positive risk (FPR)|url=http://fpr-calc.ucl.ac.uk/|publisher=UCL}}</ref> This so-called reverse Bayesian approach, which was suggested by Matthews (2001),<ref>{{cite journal|last1=Matthews|first1=R. A. J.|title=Why should clinicians care about Bayesian methods?|journal=Journal of Statistical Planning and Inference|date=2001|volume=94|pages=43–58|doi=10.1016/S0378-3758(00)00232-9}}</ref> is one way to avoid the problem that the prior probability is rarely known.
+
The hazards of reliance on ''p''-values were emphasized by pointing out that even observation of ''p'' = 0.001 was not necessarily strong evidence against the null hypothesis.<ref name="DC2017" /> Despite the fact that the likelihood ratio in favour of the alternative hypothesis over the null is close to 100, if the hypothesis was implausible, with a prior probability of a real effect being 0.1, even the observation of ''p'' = 0.001 would have a false positive risk of 8 percent. It would not even reach the 5 percent level.
 +
 
 +
It was recommended<ref name="DC2017" /> that the terms "significant" and "non-significant" should not be used. ''p''-values and confidence intervals should still be specified, but they should be accompanied by an indication of the false positive risk. It was suggested that the best way to do this is to calculate the prior probability that would be necessary to believe in order to achieve a false positive risk of, say, 5%. The calculations can be done with R-scripts that are provided,<ref name="DC2017" /> or, more simply, with a web calculator.<ref name="fpr-calc">{{cite web|last1=Longstaff|first1=Colin|last2=Colquhoun|first2=David|title=Calculator for false positive risk (FPR)|url=http://fpr-calc.ucl.ac.uk/|publisher=UCL}}</ref> This so-called reverse Bayesian approach, which was suggested by Matthews (2001),<ref>{{cite journal|last1=Matthews|first1=R. A. J.|title=Why should clinicians care about Bayesian methods?|journal=Journal of Statistical Planning and Inference|date=2001|volume=94|pages=43–58|doi=10.1016/S0378-3758(00)00232-9}}</ref> is one way to avoid the problem that the prior probability is rarely known.
  
 
=== Encouraging larger sample sizes ===
 
=== Encouraging larger sample sizes ===
To improve the quality of replications, larger [[Sample size determination|sample sizes]] than those used in the original study are often needed.<ref>{{Cite journal|title = Is psychology suffering from a replication crisis? What does "failure to replicate" really mean?|journal = American Psychologist|pages = 487–498|volume = 70|issue = 6|doi = 10.1037/a0039400|first = Scott E.|last = Maxwell|first2 = Michael Y.|last2 = Lau|first3 = George S.|last3 = Howard|pmid=26348332|year = 2015}}</ref> Larger sample sizes are needed because estimates of [[effect size]]s in published work are often exaggerated due to publication bias and large sampling variability associated with small sample sizes in an original study.<ref>{{Cite journal|title = Small studies are more heterogeneous than large ones: a meta-meta-analysis|journal = Journal of Clinical Epidemiology|pages = 860–869|volume = 68|issue = 8|doi = 10.1016/j.jclinepi.2015.03.017|pmid = 25959635|first = Joanna|last = IntHout|first2 = John P. A.|last2 = Ioannidis|first3 = George F.|last3 = Borm|first4 = Jelle J.|last4 = Goeman|year = 2015}}</ref><ref name="Button 365–376">{{Cite journal|title = Power failure: why small sample size undermines the reliability of neuroscience|journal = Nature Reviews Neuroscience|date = 2013-05-01|issn = 1471-003X|pages = 365–376|volume = 14|issue = 5|doi = 10.1038/nrn3475|first = Katherine S.|last = Button|first2 = John P. A.|last2 = Ioannidis|first3 = Claire|last3 = Mokrysz|first4 = Brian A.|last4 = Nosek|first5 = Jonathan|last5 = Flint|first6 = Emma S. J.|last6 = Robinson|first7 = Marcus R.|last7 = Munafò|pmid=23571845}}</ref><ref name="Button 365–376"/><ref>{{Cite journal|title = Consequences of prejudice against the null hypothesis|url = http://content.apa.org/journals/bul/82/1/1|journal = Psychological Bulletin|pages = 1–20|volume = 82|issue = 1|doi = 10.1037/h0076157|language = en-US|first = Anthony G.|last = Greenwald|year = 1975}}</ref> Further, using [[Statistical significance|significance thresholds]] usually leads to inflated effects, because particularly with small sample sizes, only the largest effects will become significant.<ref>{{cite journal|last1=Amrhein|first1=Valentin|last2=Korner-Nievergelt|first2=Fränzi|last3=Roth|first3=Tobias|title=The earth is flat (p > 0.05): significance thresholds and the crisis of unreplicable research|journal=PeerJ|date=2017|volume=5|page=e3544|doi=10.7717/peerj.3544|pmid=28698825|pmc=5502092}}</ref>
+
 
 +
To improve the quality of replications, larger sample sizes than those used in the original study are often needed.<ref>{{Cite journal|title = Is psychology suffering from a replication crisis? What does "failure to replicate" really mean?|journal = American Psychologist|pages = 487–498|volume = 70|issue = 6|doi = 10.1037/a0039400|first = Scott E.|last = Maxwell|first2 = Michael Y.|last2 = Lau|first3 = George S.|last3 = Howard|pmid=26348332|year = 2015}}</ref> Larger sample sizes are needed because estimates of effect sizes in published work are often exaggerated due to publication bias and large sampling variability associated with small sample sizes in an original study.<ref>{{Cite journal|title = Small studies are more heterogeneous than large ones: a meta-meta-analysis|journal = Journal of Clinical Epidemiology|pages = 860–869|volume = 68|issue = 8|doi = 10.1016/j.jclinepi.2015.03.017|pmid = 25959635|first = Joanna|last = IntHout|first2 = John P. A.|last2 = Ioannidis|first3 = George F.|last3 = Borm|first4 = Jelle J.|last4 = Goeman|year = 2015}}</ref><ref name="Button 365–376">{{Cite journal|title = Power failure: why small sample size undermines the reliability of neuroscience|journal = Nature Reviews Neuroscience|date = 2013-05-01|issn = 1471-003X|pages = 365–376|volume = 14|issue = 5|doi = 10.1038/nrn3475|first = Katherine S.|last = Button|first2 = John P. A.|last2 = Ioannidis|first3 = Claire|last3 = Mokrysz|first4 = Brian A.|last4 = Nosek|first5 = Jonathan|last5 = Flint|first6 = Emma S. J.|last6 = Robinson|first7 = Marcus R.|last7 = Munafò|pmid=23571845}}</ref><ref name="Button 365–376" /><ref>{{Cite journal|title = Consequences of prejudice against the null hypothesis|url = http://content.apa.org/journals/bul/82/1/1|journal = Psychological Bulletin|pages = 1–20|volume = 82|issue = 1|doi = 10.1037/h0076157|language = en-US|first = Anthony G.|last = Greenwald|year = 1975}}</ref> Further, using significance thresholds usually leads to inflated effects, because particularly with small sample sizes, only the largest effects will become significant.<ref>{{cite journal|last1=Amrhein|first1=Valentin|last2=Korner-Nievergelt|first2=Fränzi|last3=Roth|first3=Tobias|title=The earth is flat (p > 0.05): significance thresholds and the crisis of unreplicable research|journal=PeerJ|date=2017|volume=5|page=e3544|doi=10.7717/peerj.3544|pmid=28698825|pmc=5502092}}</ref>
  
 
=== Sharing raw data in online repositories ===
 
=== Sharing raw data in online repositories ===
Online repositories where data, protocols, and findings can be stored and evaluated by the public seek to improve the integrity and reproducibility of research. Examples of such repositories include the [[Center for Open Science#Open Science Framework|Open Science Framework]], [[Registry of Research Data Repositories]], and Psychfiledrawer.org. Sites like Open Science Framework offer badges for using open science practices in an effort to incentivize scientists. However, there has been concern that those who are most likely to provide their data and code for analyses are the researchers that are likely the most sophisticated.<ref name=":2">{{Cite journal|title = Anticipating consequences of sharing raw data and code and of awarding badges for sharing|journal = Journal of Clinical Epidemiology|volume = 70|pages = 258–260|doi = 10.1016/j.jclinepi.2015.04.015|pmid = 26163123|first = John P. A.|last = Ioannidis|year = 2016}}</ref> John Ioannidis at Stanford University suggested that "the paradox may arise that the most meticulous and sophisticated and method-savvy and careful researchers may become more susceptible to criticism and reputation attacks by reanalyzers who hunt for errors, no matter how negligible these errors are".<ref name=":2" />
+
 
 +
Online repositories where data, protocols, and findings can be stored and evaluated by the public seek to improve the integrity and reproducibility of research. Examples of such repositories include the Open Science Framework, Registry of Research Data Repositories, and Psychfiledrawer.org. Sites like Open Science Framework offer badges for using open science practices in an effort to incentivize scientists. However, there has been concern that those who are most likely to provide their data and code for analyses are the researchers that are likely the most sophisticated anyway.<ref name=":2">{{Cite journal|title = Anticipating consequences of sharing raw data and code and of awarding badges for sharing|journal = Journal of Clinical Epidemiology|volume = 70|pages = 258–260|doi = 10.1016/j.jclinepi.2015.04.015|pmid = 26163123|first = John P. A.|last = Ioannidis|year = 2016}}</ref> John Ioannidis at Stanford University suggested that "the paradox may arise that the most meticulous and sophisticated and method-savvy and careful researchers may become more susceptible to criticism and reputation attacks by reanalyzers who hunt for errors, no matter how negligible these errors are".<ref name=":2" />
  
 
=== Funding for replication studies ===
 
=== Funding for replication studies ===
In July 2016 the [[Netherlands Organisation for Scientific Research]] made €3 million available for replication studies. The funding is for replication based on reanalysis of existing data and replication by collecting and analysing new data. Funding is available in the areas of social sciences, health research and healthcare innovation.<ref>{{cite web|title=NWO makes 3 million available for Replication Studies pilot|url=http://www.nwo.nl/en/news-and-events/news/2016/nwo-makes-3-million-available-for-replication-studies-pilot.html|website=NWO|accessdate=2 August 2016}}</ref>
 
  
In 2013 the [[Laura and John Arnold Foundation]] funded the launch of [[Reproducibility Project#Center for Open Science|The Center for Open Science]] with a $5.25 million grant and by 2017 had provided an additional $10 million in funding.<ref name=Wired2017>{{cite news|last1=Apple|first1=Sam|title=The Young Billionaire Behind the War on Bad Science|url=https://www.wired.com/2017/01/john-arnold-waging-war-on-bad-science/|work=Wired|date=January 22, 2017}}</ref> It also funded the launch of the [[Meta-Research Innovation Center at Stanford]] at Stanford University run by [[John Ioannidis]] and Steven Goodman to study ways to improve scientific research.<ref name=Wired2017/> It also provided funding for the [[AllTrials]] initiative led in part by [[Ben Goldacre]].<ref name=Wired2017/>
+
In July 2016 the Netherlands Organisation for Scientific Research made €3 million available for replication studies. The funding is for replication based on reanalysis of existing data and replication by collecting and analysing new data. Funding is available in the areas of social sciences, health research and healthcare innovation.<ref>{{cite web|title=NWO makes 3 million available for Replication Studies pilot|url=http://www.nwo.nl/en/news-and-events/news/2016/nwo-makes-3-million-available-for-replication-studies-pilot.html|website=NWO|accessdate=2 August 2016}}</ref>
 +
 
 +
In 2013 the Laura and John Arnold Foundation funded the launch of The Center for Open Science with a $5.25 million grant and by 2017 had provided an additional $10 million in funding.<ref name="Wired2017">{{cite news|last1=Apple|first1=Sam|title=The Young Billionaire Behind the War on Bad Science|url=https://www.wired.com/2017/01/john-arnold-waging-war-on-bad-science/|work=Wired|date=January 22, 2017}}</ref> It also funded the launch of the Meta-Research Innovation Center at Stanford at Stanford University run by [[John_Ioannidis|John Ioannidis]] and Steven Goodman to study ways to improve scientific research.<ref name="Wired2017" /> It also provided funding for the AllTrials initiative led in part by Ben Goldacre.<ref name="Wired2017" />
  
 
=== Emphasize triangulation, not just replication ===
 
=== Emphasize triangulation, not just replication ===
  
Marcus R. Munafò and George Davey Smith argue, in a piece published by ''[[Nature (journal)|Nature]]'', that research should emphasize triangulation, not just replication. They claim that,
+
Marcus R. Munafò and George Davey Smith argue, in a piece published by ''Nature'', that research should emphasize triangulation, not just replication. They claim that,
  
{{quote|replication alone will get us only so far (and) might actually make matters worse&nbsp;... We believe that an essential protection against flawed ideas is triangulation. This is the strategic use of multiple approaches to address one question. Each approach has its own unrelated assumptions, strengths and weaknesses. Results that agree across different methodologies are less likely to be [[Artifact (error)|artefacts]].&nbsp;... Maybe one reason replication has captured so much interest is the often-repeated idea that falsification is at the heart of the scientific enterprise. This idea was popularized by [[Karl Popper]]'s 1950s maxim that theories can never be proved, only [[Falsifiability|falsified]]. Yet an overemphasis on repeating experiments could provide an unfounded sense of certainty about findings that rely on a single approach.&nbsp;... philosophers of science have moved on since Popper. Better descriptions of how scientists actually work include what epistemologist [[Peter Lipton]] called in 1991 "inference to the best explanation".<ref name=Munafo2018>{{cite journal|last1=Munafò|first1=Marcus R.|last2=Smith|first2=George Davey|title=Robust research needs many lines of evidence|journal=Nature|volume=553|issue=7689|pages=399–401|date=January 23, 2018|doi=10.1038/d41586-018-01023-3|pmid=29368721|bibcode=2018Natur.553..399M}}</ref>}}
+
<blockquote>replication alone will get us only so far (and) might actually make matters worse&nbsp;... We believe that an essential protection against flawed ideas is triangulation. This is the strategic use of multiple approaches to address one question. Each approach has its own unrelated assumptions, strengths and weaknesses. Results that agree across different methodologies are less likely to be artefacts ... Maybe one reason replication has captured so much interest is the often-repeated idea that falsification is at the heart of the scientific enterprise. This idea was popularized by [[Karl Popper]]'s 1950s maxim that theories can never be proved, only falsified. Yet an overemphasis on repeating experiments could provide an unfounded sense of certainty about findings that rely on a single approach.&nbsp;... philosophers of science have moved on since Popper. Better descriptions of how scientists actually work include what epistemologist Peter Lipton called in 1991 "inference to the best explanation".<ref name=Munafo2018>{{cite journal|last1=Munafò|first1=Marcus R.|last2=Smith|first2=George Davey|title=Robust research needs many lines of evidence|journal=Nature|volume=553|issue=7689|pages=399–401|date=January 23, 2018|doi=10.1038/d41586-018-01023-3|pmid=29368721|bibcode=2018Natur.553..399M}}</ref></blockquote>
 +
 
 +
== Further reading ==
 +
 
 +
*{{cite web |last1=Bastian |first1=Hilda |title=Reproducibility Crisis Timeline: Milestones in Tackling Research Reliability |url=https://blogs.plos.org/absolutely-maybe/2016/12/05/reproducibility-crisis-timeline-milestones-in-tackling-research-reliability/ |website=Absolutely Maybe |accessdate=5 June 2019 |date=5 December 2016}}
 +
*{{cite book |last1=Harris |first1=Richard |title=Rigor Mortis: How Sloppy Science Creates Worthless Cures, Crushes Hope, and Wastes Billions |date=2017 |publisher=Basic Books |location=New York |isbn=9780465097906}}  
 +
*[https://www.firstthings.com/article/2016/05/scientific-regress Scientific Regress] – William A. Wilson in First Things, May 2016
 +
*[https://www.displayr.com/what-is-the-replication-crisis/ What is the Replication Crisis?] by Tim Bock. – Being a short summary of what the replication crisis is about.
 +
*[http://www.plosmedicine.org/article/info:doi/10.1371/journal.pmed.0020124 Why Most Published Research Findings Are False] by John P. A. Ioannidis, Public Library of Science Medicine, 30 August 2005.
 +
*[https://fabiusmaximus.com/2016/04/19/replication-crisis-in-science-95394/ The replication crisis in science has just begun. It will be big.] by Larry Kummer, Editor Science & Nature 19 April 2016 on Fabius Maximus website.
  
 
== References ==
 
== References ==
 +
 
<references />
 
<references />
 
== Further reading ==
 
* [http://journals.plos.org/plosmedicine/article?id=10.1371/journal.pmed.0020124 Why Most Published Research Findings Are False]
 
* {{cite web |last1=Bastian |first1=Hilda |title=Reproducibility Crisis Timeline: Milestones in Tackling Research Reliability |url=https://blogs.plos.org/absolutely-maybe/2016/12/05/reproducibility-crisis-timeline-milestones-in-tackling-research-reliability/ |website=Absolutely Maybe |accessdate=5 June 2019 |date=5 December 2016}}
 
* {{cite book |last1=Harris |first1=Richard |title=Rigor Mortis: How Sloppy Science Creates Worthless Cures, Crushes Hope, and Wastes Billions |date=2017 |publisher=Basic Books |location=New York |isbn=9780465097906}}
 

Latest revision as of 17:36, 8 September 2019

Parts of the following article are based on the Wikipedia article „Replication Crisis“ from Wikipedia, as read on 31.8.2019, under the licence of Creative Commons CC-BY-SA 3.0 Unported (short version). A list of the authors is available on the respective page of Wikipedia. Changes are possible and probable.

This article is closely connected to the article on the Decline Effect, which is part of this crisis. For further explanation and material please see also there.

The replication crisis (or replicability crisis or reproducibility crisis) is an ongoing (2019) methodological crisis in which it has been found that many scientific studies are difficult or impossible to replicate or reproduce. The replication crisis affects the social and life sciences most severely,[1][2] whereas the hard sciences are still reluctant to face the problem in the first place and mainly don’t even use double blinding in their research.[3] The crisis has long-standing roots. The phrase was coined in the early 2010s[4] as part of a growing awareness of the problem. The replication crisis represents an important body of research in metascience.[5]

Because the reproducibility of experiments is an essential part of the scientific method,[6] the inability to replicate the studies of others has potentially grave consequences for many fields of science in which significant theories are grounded on unreproducible experimental work. The replication crisis has been particularly widely discussed in the field of psychology (and in particular, social psychology) and in medicine, where a number of efforts have been made to re-investigate classic results, and to attempt to determine both the reliability of the results, and, if found to be unreliable, the reasons for the failure of replication.[7][8]

Scope of the crisis

Overall

According to a 2016 poll of 1,500 scientists reported that 70% of them had failed to reproduce at least one other scientist's experiment (50% had failed to reproduce one of their own experiments).[9] In 2009, 2% of scientists admitted to falsifying studies at least once and 14% admitted to personally knowing someone who did. Misconducts were reported more frequently by medical researchers than others.[10]

Psychology

Several factors have combined to put psychology at the center of the controversy.[11] Much of the focus has been on the area of social psychology,[12] although other areas of psychology such as clinical psychology,[13][14] developmental psychology,[15] and educational research have also been implicated.[16][17] According to a 2018 survey of 200 meta-analyses, "psychological research is, on average, afflicted with low statistical power".[18]

Firstly, questionable research practices (QRPs) have been identified as common in the field.[19] Such practices, while not intentionally fraudulent, involve capitalizing on the gray area of acceptable scientific practices or exploiting flexibility in data collection, analysis, and reporting, often in an effort to obtain a desired outcome. Examples of QRPs include selective reporting or partial publication of data (reporting only some of the study conditions or collected dependent measures in a publication), optional stopping (choosing when to stop data collection, often based on statistical significance of tests), p-value rounding (rounding p-values down to 0.05 to suggest statistical significance), file drawer effect (nonpublication of data), post-hoc storytelling (framing exploratory analyses as confirmatory analyses), and manipulation of outliers (either removing outliers or leaving outliers in a dataset to cause a statistical test to be significant).[19][20][21][22] A survey of over 2,000 psychologists indicated that a majority of respondents admitted to using at least one QRP.[19] False positive conclusions, often resulting from the pressure to publish or the author's own confirmation bias, are an inherent hazard in the field, requiring a certain degree of skepticism on the part of readers.[23]

Secondly, psychology and social psychology in particular, has found itself at the center of several scandals involving outright fraudulent research, most notably the admitted data fabrication by Diederik Stapel[24] as well as allegations against others. However, most scholars acknowledge that fraud is, perhaps, the lesser contribution to replication crises.

Thirdly, several effects in psychological science have been found to be difficult to replicate even before the current replication crisis. For example, the scientific journal Judgment and Decision Making has published several studies over the years that fail to provide support for the unconscious thought theory. Replications appear particularly difficult when research trials are pre-registered and conducted by research groups not highly invested in the theory under questioning.

These three elements together have resulted in renewed attention for replication supported by psychologist Daniel Kahneman.[25] Scrutiny of many effects have shown that several core beliefs are hard to replicate. A recent special edition of the journal Social Psychology focused on replication studies and a number of previously held beliefs were found to be difficult to replicate.[26] A 2012 special edition of the journal Perspectives on Psychological Science also focused on issues ranging from publication bias to null-aversion that contribute to the replication crises in psychology.[27] In 2015, the first open empirical study of reproducibility in Psychology was published, called the Reproducibility Project. Researchers from around the world collaborated to replicate 100 empirical studies from three top Psychology journals. Fewer than half of the attempted replications were successful at producing statistically significant results in the expected directions, though most of the attempted replications did produce trends in the expected directions.[28]

Many research trials and meta-analyses are compromised by poor quality and conflicts of interest that involve both authors and professional advocacy organizations, resulting in many false positives regarding the effectiveness of certain types of psychotherapy.[29]

Although the British newspaper The Independent wrote that the results of the reproducibility project show that much of the published research is just "psycho-babble",[30] the replication crisis does not necessarily mean that psychology is unscientific.[31][32][33] Rather this process is a healthy if sometimes acrimonious part of the scientific process in which old ideas or those that cannot withstand careful scrutiny are pruned,[34][35] although this pruning process is not always effective.[36][37] The consequence is that some areas of psychology once considered solid, such as social priming, have come under increased scrutiny due to failed replications.[38]

Nobel laureate and professor emeritus in psychology Daniel Kahneman argued that the original authors should be involved in the replication effort because the published methods are often too vague.[39] Others such as Dr. Andrew Wilson disagree and argue that the methods should be written down in detail.[39] An investigation of replication rates in psychology in 2012 indicated higher success rates of replication in replication studies when there was author overlap with the original authors of a study[40] (91.7% successful replication rates in studies with author overlap compared to 64.6% success replication rates without author overlap).

Psychology replication rates

A report by the Open Science Collaboration in August 2015 that was coordinated by Brian Nosek estimated the reproducibility of 100 studies in psychological science from three high-ranking psychology journals.[41] Overall, 36% of the replications yielded significant findings (p value below 0.05) compared to 97% of the original studies that had significant effects. The mean effect size in the replications was approximately half the magnitude of the effects reported in the original studies.

The same paper examined the reproducibility rates and effect sizes by journal (Journal of Personality and Social Psychology [JPSP], Journal of Experimental Psychology: Learning, Memory, and Cognition [JEP:LMC], Psychological Science [PSCI]) and discipline (social psychology, cognitive psychology). Study replication rates were 23% for JPSP, 38% for JEP:LMC, and 38% for PSCI. Studies in the field of cognitive psychology had a higher replication rate (50%) than studies in the field of social psychology (25%).

An analysis of the publication history in the top 100 psychology journals between 1900 and 2012 indicated that approximately 1.6% of all psychology publications were replication attempts.[40] Articles were considered a replication attempt if the term "replication" appeared in the text. A subset of those studies (500 studies) was randomly selected for further examination and yielded a lower replication rate of 1.07% (342 of the 500 studies [68.4%] were actually replications). In the subset of 500 studies, analysis indicated that 78.9% of published replication attempts were successful. The rate of successful replication was significantly higher when at least one author of the original study was part of the replication attempt (91.7% relative to 64.6%).

A study published in 2018 in Nature Human Behaviour sought to replicate 21 social and behavioral science papers from Nature and Science, finding that only 13 could be successfully replicated.[42][43] Similarly, in a study conducted under the auspices of the Center for Open Science, a team of 186 researchers from 60 different laboratories (representing 36 different nationalities from 6 different continents) conducted replications of 28 classic and contemporary findings in psychology.[44] The focus of the study was not only on whether or not the findings from the original papers replicated, but also on the extent to which findings varied as a function of variations in samples and contexts. Overall, 14 of the 28 findings failed to replicate despite massive sample sizes. However, if a finding replicated, it replicated in most samples, while if a finding was not replicated, it failed to replicate with little variation across samples and contexts. This evidence is inconsistent with a popular explanation that failures to replicate in psychology are likely due to changes in the sample between the original and replication study.[45]

A disciplinary social dilemma

Highlighting the social structure that discourages replication in psychology, Brian D. Earp and Jim A. C. Everett enumerated five points as to why replication attempts are uncommon:[46][47]

  1. "Independent, direct replications of others' findings can be time-consuming for the replicating researcher
  2. "[Replications] are likely to take energy and resources directly away from other projects that reflect one's own original thinking
  3. "[Replications] are generally harder to publish (in large part because they are viewed as being unoriginal)
  4. "Even if [replications] are published, they are likely to be seen as 'bricklaying' exercises, rather than as major contributions to the field
  5. "[Replications] bring less recognition and reward, and even basic career security, to their authors"[48]

For these reasons the authors advocated that psychology is facing a disciplinary social dilemma, where the interests of the discipline are at odds with the interests of the individual researcher.

Medicine

Medical researchers were among the first ones to ring alarm regarding the lack of replicability of pre-clinical studies on drugs targeted for industrial use.[49] This lack of reproducability and therefore of reliability in research started to become a growing problem for pharmaceutical companies because their decisions in which drug targets to invest money for expensive clinical research depends mainly on such studies.

Out of 49 medical studies from 1990–2003, with more than 1000 citations, 45 claimed that studied therapy was effective. Out of these studies, 16% were contradicted by subsequent studies, 16% had found stronger effects than did subsequent studies, 44% were replicated, and 24% remained largely unchallenged.[50] The US Food and Drug Administration in 1977–1990 found flaws in 10–20% of medical studies.[51] In a paper published in 2012, Glenn Begley, a biotech consultant working at Amgen, and Lee Ellis, at the University of Texas, argued that only 11% of the pre-clinical cancer studies could be replicated.[52][53]

A 2016 article by John Ioannidis, Professor of Medicine and of Health Research and Policy at Stanford University School of Medicine and a Professor of Statistics at Stanford University School of Humanities and Sciences, elaborated on "Why Most Clinical Research Is Not Useful".[54] In the article Ioannidis laid out some of the problems and called for reform, characterizing certain points for medical research to be useful again; one example he made was the need for medicine to be "patient centered" (e.g. in the form of the Patient-Centered Outcomes Research Institute) instead of the current practice to mainly take care of "the needs of physicians, investigators, or sponsors". Ioannidis is known for his research focus on science itself since the 2005 paper "Why Most Published Research Findings Are False".[55]

Marketing

Marketing is another discipline with a "desperate need" for replication.[56] Many famous marketing studies fail to be repeated upon replication, a notable example being the "too-many-choices" effect, in which a high number of choices of product makes a consumer less likely to purchase.[57] In addition to the previously mentioned arguments, replications studies in marketing are needed to examine the applicability of theories and models across countries and cultures, which is especially important because of possible influences of globalization.[58]

Economics

A 2016 study in the journal Science found that one-third of 18 experimental studies from two top-tier economics journals (American Economic Review and the Quarterly Journal of Economics) failed to be successfully replicated.[59][60] A 2017 study in the Economic Journal suggested that "the majority of the average effects in the empirical economics literature are exaggerated by a factor of at least 2 and at least one-third are exaggerated by a factor of 4 or more".[61]

Sports Science

A 2018 study took the field of exercise and sports science to task for insufficient replication studies, limited reporting of null results and trivial results, and insufficient research transparency.[62] Statisticians have criticized sports science for common use of a controversial statistical method called "magnitude-based inference" that has allowed sports scientists to extract apparently significant results from noisy data where ordinary hypothesis testing would have found none.[63]

Hydrology

A 2019 study in Scientific Data found that a small number of articles in hydrology and water resources journals could be reproduced due to data unavailability. The study "estimated, with 95% confidence, that results might be reproduced for only 0.6% to 6.8% of all 1,989 articles".[64][65][66]
This study mainly relates to the quality of documentation of origninal material which is obviously only given scarcely or not at all. With no raw data available the factual validity of the scientific articles cannot be estimated, but the scientific non-reproducability is not proven thereby.

Hard Sciences

As opposed to the social and life sciences many agree that the “hard sciences” – physics, chemistry, biology, astronomy, etc. – will not be touched by this crisis. Nevertheless there seem to be quite a few problems there as well. „Even physics has been affected, as William Wilson notes. “Two of the most vaunted physics results of the past few years — the announced discovery of both cosmic inflation and gravitational waves at the BICEP2 experiment in Antarctica, and the supposed discovery of superluminal neutrinos at the Swiss-Italian border — have now been retracted, with far less fanfare than when they were first published.” See this about the former and this about the latter.”[67]

Natural constants

One of the pillars of physics and all sciences based on the laws described in physics is the assumption of the constancy of the natural constants. This constancy is a general assumption because it can principally not been proven. One can only state that they have been constant since humans are able to measure them, which is an extremely short time span compared to the age of the universe.
Nevertheless there is good reason to doubt even this basic assumption of the physical sciences.„At the end of 1998 the CODATA even decided to increase the uncertainty of the accepted value for the gravitational constant from 128 ppm to 1500 ppm. This remarkable step of increasing the uncertainty instead of decreasing was made to reflect the discrepancies between recent experiments, which span a wide range of more than 0.7 %.”[68]
And in his work “The Science Delusion – Freeing the Spirit of Enquiry” biologist and philosopher Rupert Sheldrake[69] discusses this problem and gives lots of examples showing that the fundamental constants may not be as eternal as was thought but object to change by time. He prefers to see the “constants” more as habits of nature than as eternal laws. Sheldrake points out that the fundamental physical constants are artificially held as constant by defining them to be so and by reducing all measurements to the mean and eliminating strongly deviating measurings from the count. So as a matter of fact scientific measurings of the constants constantly produce quite different results, which are collected and regulated by the Committee on Data for Science and Technology (CODATA) to always be constant as by definition.[70]

Causes of the crisis

Bad Science

The Sokal affair and the discussion in its wake has shown clearly that there is a major problem of not distinguishing between real and fake science.[71]
The editor of the prominent medical journal Lancet[72] Richard Horton writes about “apparent endemicity of bad research behaviour” and “bad scientific practices” when he points out that “much of the scientific literature, perhaps half, may simply be untrue.” As the editor of the probably most important medical journal he is in a position to recognize the problem.

High rate of publications

In fact some predictions of a possible crisis in the quality control mechanism of science can be traced back several decades, especially among scholars in science and technology studies (STS). Derek de Solla Price – considered the father of scientometrics – predicted that science could reach 'senility' as a result of its own exponential growth.[73] Some present day literature seems to vindicate this 'overflow' prophesy, lamenting at decay in both attention and quality.[74][75]

Philosopher and historian of science Jerome R. Ravetz predicted in his 1971 book Scientific Knowledge and Its Social Problems that science – in moving from the little science made of restricted communities of scientists to big science or techno-science – would suffer major problems in its internal system of quality control. Ravetz anticipated that modern science's system of rewarding scientists for research might become dysfunctional, the present 'publish or perish' challenge, creating perverse incentives to publish any findings however dubious. For Ravetz quality in science is maintained when there is a community of scholars linked by norms and standards, and a willingness to stand by these.

Historian Philip Mirowski offered more recently a similar diagnosis in his 2011 book Science Mart (2011).[76] 'Mart' is here a reference to the retail giant 'Walmart' and an allusion to the commodification of science. In the analysis of Mirowski, when science becomes a commodity being traded in a market, its quality collapses. Mirowski argues his case by tracing the decay of science to the decision of major corporations to close their in-house laboratories in order to outsource their work to universities, and subsequently to move their research away from universities to even cheaper contract research organizations (CRO).

Insufficient control

The crisis of science's quality control system is affecting the use of science for policy. This is the thesis of a recent work by a group of STS scholars, who identify in 'evidence based (or informed) policy' a point of present tension.[77][78][79][80] Economist Noah Smith suggests that a factor in the crisis has been the overvaluing of research in academia and undervaluing of teaching ability, especially in fields with few major recent discoveries.[81] In the journal Science article “Who’s Afraid of Peer Review?”[82] author John Bohannon explains the problem of the growing number of open access journals and there lack of effective peer review. Of 300 fake articles filled with obvious mistakes and nonsense sent out to as many different online journals more than half were accepted without relevant criticism.

Publication bias

Another major cause is publication bias, the fact that positive results are more likely to get published than negative (or null) results. This may lead to the canonization of false facts [83].

Fundamental causes in the scientific paradigm

All the above reasons for lack of replicability have one thing in common: They are consequences of human fault, misconduct or mistakes and could theoretically be corrected by discipline, correctness and better control. Nevertheless there remains the possibility that – apart from all those considerations and observation being correct – there could also be more fundamental problems of the axiom of replicability and the constancy of nature as such.
The biologist, philosopher and meta-scientist Rupert Sheldrake has pointed out that there is ample reason to doubt the belief in the full replicability and reliability of the hard sciences as well and that there might be a problem with the assumption of replicability as such.[84]

In any case, the Decline effect (see main article) indicates general problems of replicability because it cannot be fully explained by the causes proposed so far. It was discovered by J.B.Rhine in the 20s of the last century, long before the current crisis and under completely different circumstances. Rhine found the effect in his own research, even in areas whose statistical validity he himself had previously verified.

Public Response

Political Repercussions

In the US, science's reproducibility crisis has become a topic of political contention, linked to the attempt to diminish regulations – e.g. of emissions of pollutants, with the argument that these regulations are based on non-reproducible science.[85][80] Previous attempts with the same aim accused studies of being non-transparent.[86]

Public discussion

Apart from studies and articles in scientific journals there has been little public discussion of the Replicability Crisis as yet and few reports in the mass media. Most of the discussions are related to statistical, economical and sociological reasons only.[87]
The question whether there is a general problem of replicability behind these findings, too, has not been tackled so far. And there has been very little effort in the hard sciences to deal with the replication issue as there seems to be a consense that the problem is only one of the social and life sciences.

Addressing the replication crisis

Replication has been referred to as "the cornerstone of science".[88][89] Replication studies attempt to evaluate whether published results reflect true findings or false positives. The integrity of scientific findings and reproducibility of research are important as they form the knowledge foundation on which future studies are built.

Metascience

Metascience is the use of scientific methodology to study science itself. Metascience seeks to increase the quality of scientific research while reducing waste. It is also known as "research on research" and "the science of science", as it uses research methods to study how research is done and where improvements can be made. Metascience concerns itself with all fields of research and has been described as "a bird's eye view of science."[90] In the words of John Ioannidis, "Science is the best thing that has happened to human beings ... but we can do it better."[91]

Meta-research continues to be conducted to identify the roots of the crisis and to address them. Methods of addressing the crisis include pre-registration of scientific studies and clinical trials as well as the founding of organizations such as CONSORT and the EQUATOR Network that issue guidelines for methodology and reporting. There are continuing efforts to reform the system of academic incentives, to improve the peer review process, to reduce the misuse of statistics, to combat bias in scientific literature, and to increase the overall quality and efficiency of the scientific process.

Tackling publication bias with pre-registration of studies

A recent innovation in scientific publishing to address the replication crisis is through the use of registered reports.[92][93] The registered report format requires authors to submit a description of the study methods and analyses prior to data collection. Once the method and analysis plan is vetted through peer-review, publication of the findings is provisionally guaranteed, based on whether the authors follow the proposed protocol. One goal of registered reports is to circumvent the publication bias toward significant findings that can lead to implementation of questionable research practices and to encourage publication of studies with rigorous methods.

The journal Psychological Science has encouraged the preregistration of studies and the reporting of effect sizes and confidence intervals.[94] The editor in chief also noted that the editorial staff will be asking for replication of studies with surprising findings from examinations using small sample sizes before allowing the manuscripts to be published.

Moreover, only a very small proportion of academic journals in psychology and neurosciences explicitly stated that they welcome submissions of replication studies in their aim and scope or instructions to authors.[95][96] This phenomenon does not encourage the reporting or even attempt on replication studies.

Emphasizing replication attempts in teaching

Based on coursework in experimental methods at MIT and Stanford, it has been suggested that methods courses in psychology emphasize replication attempts rather than original studies.[97][98] Such an approach would help students learn scientific methodology and provide numerous independent replications of meaningful scientific findings that would test the replicability of scientific findings. Some have recommended that graduate students should be required to publish a high-quality replication attempt on a topic related to their doctoral research prior to graduation.[47]

Reducing the p-value required for claiming significance of new results

Many publications require a p-value of p < 0.05 to claim statistical significance. The paper "Redefine statistical significance",[99] signed by a large number of scientists and mathematicians, proposes that in "fields where the threshold for defining statistical significance for new discoveries is P < 0.05, we propose a change to P < 0.005. This simple step would immediately improve the reproducibility of scientific research in many fields."

Their rationale is that "a leading cause of non-reproducibility (is that the) statistical standards of evidence for claiming new discoveries in many fields of science are simply too low. Associating 'statistically significant' findings with P < 0.05 results in a high rate of false positives even in the absence of other experimental, procedural and reporting problems."

Addressing the misinterpretation of p-values

Although statisticians are unanimous that use of the p < 0.05 provides weaker evidence than is generally appreciated, there is a lack of unanimity about what should be done about it. Some have advocated that Bayesian methods should replace p-values. This has not happened on a wide scale, partly because it is complicated, and partly because many users distrust the specification of prior distributions in the absence of hard data. A simplified version of the Bayesian argument, based on testing a point null hypothesis was suggested by Colquhoun (2014, 2017).[100][101] The logical problems of inductive inference were discussed in "The problem with p-values" (2016).[102]

The hazards of reliance on p-values were emphasized by pointing out that even observation of p = 0.001 was not necessarily strong evidence against the null hypothesis.[101] Despite the fact that the likelihood ratio in favour of the alternative hypothesis over the null is close to 100, if the hypothesis was implausible, with a prior probability of a real effect being 0.1, even the observation of p = 0.001 would have a false positive risk of 8 percent. It would not even reach the 5 percent level.

It was recommended[101] that the terms "significant" and "non-significant" should not be used. p-values and confidence intervals should still be specified, but they should be accompanied by an indication of the false positive risk. It was suggested that the best way to do this is to calculate the prior probability that would be necessary to believe in order to achieve a false positive risk of, say, 5%. The calculations can be done with R-scripts that are provided,[101] or, more simply, with a web calculator.[103] This so-called reverse Bayesian approach, which was suggested by Matthews (2001),[104] is one way to avoid the problem that the prior probability is rarely known.

Encouraging larger sample sizes

To improve the quality of replications, larger sample sizes than those used in the original study are often needed.[105] Larger sample sizes are needed because estimates of effect sizes in published work are often exaggerated due to publication bias and large sampling variability associated with small sample sizes in an original study.[106][107][107][108] Further, using significance thresholds usually leads to inflated effects, because particularly with small sample sizes, only the largest effects will become significant.[109]

Sharing raw data in online repositories

Online repositories where data, protocols, and findings can be stored and evaluated by the public seek to improve the integrity and reproducibility of research. Examples of such repositories include the Open Science Framework, Registry of Research Data Repositories, and Psychfiledrawer.org. Sites like Open Science Framework offer badges for using open science practices in an effort to incentivize scientists. However, there has been concern that those who are most likely to provide their data and code for analyses are the researchers that are likely the most sophisticated anyway.[110] John Ioannidis at Stanford University suggested that "the paradox may arise that the most meticulous and sophisticated and method-savvy and careful researchers may become more susceptible to criticism and reputation attacks by reanalyzers who hunt for errors, no matter how negligible these errors are".[110]

Funding for replication studies

In July 2016 the Netherlands Organisation for Scientific Research made €3 million available for replication studies. The funding is for replication based on reanalysis of existing data and replication by collecting and analysing new data. Funding is available in the areas of social sciences, health research and healthcare innovation.[111]

In 2013 the Laura and John Arnold Foundation funded the launch of The Center for Open Science with a $5.25 million grant and by 2017 had provided an additional $10 million in funding.[112] It also funded the launch of the Meta-Research Innovation Center at Stanford at Stanford University run by John Ioannidis and Steven Goodman to study ways to improve scientific research.[112] It also provided funding for the AllTrials initiative led in part by Ben Goldacre.[112]

Emphasize triangulation, not just replication

Marcus R. Munafò and George Davey Smith argue, in a piece published by Nature, that research should emphasize triangulation, not just replication. They claim that,

replication alone will get us only so far (and) might actually make matters worse ... We believe that an essential protection against flawed ideas is triangulation. This is the strategic use of multiple approaches to address one question. Each approach has its own unrelated assumptions, strengths and weaknesses. Results that agree across different methodologies are less likely to be artefacts ... Maybe one reason replication has captured so much interest is the often-repeated idea that falsification is at the heart of the scientific enterprise. This idea was popularized by Karl Popper's 1950s maxim that theories can never be proved, only falsified. Yet an overemphasis on repeating experiments could provide an unfounded sense of certainty about findings that rely on a single approach. ... philosophers of science have moved on since Popper. Better descriptions of how scientists actually work include what epistemologist Peter Lipton called in 1991 "inference to the best explanation".[113]

Further reading

  • Reproducibility Crisis Timeline: Milestones in Tackling Research Reliability (5 December 2016). Retrieved on 5 June 2019.
  • Harris, Richard (2017). Rigor Mortis: How Sloppy Science Creates Worthless Cures, Crushes Hope, and Wastes Billions. New York: Basic Books. ISBN 9780465097906.
  • Scientific Regress – William A. Wilson in First Things, May 2016
  • What is the Replication Crisis? by Tim Bock. – Being a short summary of what the replication crisis is about.
  • Why Most Published Research Findings Are False by John P. A. Ioannidis, Public Library of Science Medicine, 30 August 2005.
  • The replication crisis in science has just begun. It will be big. by Larry Kummer, Editor Science & Nature 19 April 2016 on Fabius Maximus website.

References

  1. Schooler, J. W. (2014). "Metascience could rescue the 'replication crisis'". Nature. 515 (7525): 9. Bibcode:2014Natur.515....9S. doi:10.1038/515009a. PMID 25373639.
  2. Why 'Statistical Significance' Is Often Insignificant. Retrieved on 7 November 2017.
  3. Rupert Sheldrake: The Science Delusion – Freeing the Spirit of Enquiry; London 2012, Hodder& Stoughton, ISBN 978 1 444 72795 1. Chapter 11. Illusions of Objectivity.
  4. Pashler, Harold; Wagenmakers, Eric Jan (2012). "Editors' Introduction to the Special Section on Replicability in Psychological Science: A Crisis of Confidence?". Perspectives on Psychological Science. 7 (6): 528–530. doi:10.1177/1745691612465253. PMID 26168108.
  5. Reproducibility of Scientific Results. Metaphysics Research Lab, Stanford University (2018). Retrieved on 19 May 2019.
  6. Staddon, John (2017) Scientific Method: How science works, fails to work or pretends to work. Taylor and Francis.
  7. Gary Marcus (May 1, 2013). The Crisis in Social Psychology That Isn't. The New Yorker.
  8. Jonah Lehrer (December 13, 2010). The Truth Wears Off. The New Yorker.
  9. Is There a Reproducibility Crisis in Science?. Nature Video, Scientific American (28 May 2016). Retrieved on 15 August 2019.
  10. Fanelli, Daniele (29 May 2009). "How Many Scientists Fabricate and Falsify Research? A Systematic Review and Meta-Analysis of Survey Data". PLOS ONE. 4 (5): e5738. Bibcode:2009PLoSO...4.5738F. doi:10.1371/journal.pone.0005738. PMC 2685008. PMID 19478950.
  11. No, science's reproducibility problem is not limited to psychology. Retrieved on 10 September 2015.
  12. Dominus, Susan (2017-10-18). "When the Revolution Came for Amy Cuddy". The New York Times. ISSN 0362-4331. Retrieved 2017-10-19.
  13. Leichsenring, Falk; Abbass, Allan; Hilsenroth, Mark J.; Leweke, Frank; Luyten, Patrick; Keefe, Jack R.; Midgley, Nick; Rabung, Sven; Salzer, Simone; Steiner, Christiane (April 2017). "Biases in research: risk factors for non-replicability in psychotherapy and pharmacotherapy research". Psychological Medicine. 47 (6): 1000–1011. doi:10.1017/S003329171600324X. PMID 27955715.
  14. Hengartner, Michael P. (February 28, 2018). "Raising Awareness for the Replication Crisis in Clinical Psychology by Focusing on Inconsistencies in Psychotherapy Research: How Much Can We Rely on Published Findings from Efficacy Trials?". Frontiers in Psychology. Frontiers Media. 9: 256. doi:10.3389/fpsyg.2018.00256. PMC 5835722. PMID 29541051.
  15. Frank, Michael C.; Bergelson, Elika; Bergmann, Christina; Cristia, Alejandrina; Floccia, Caroline; Gervain, Judit; Hamlin, J. Kiley; Hannon, Erin E.; Kline, Melissa; Levelt, Claartje; Lew-Williams, Casey; Nazzi, Thierry; Panneton, Robin; Rabagliati, Hugh; Soderstrom, Melanie; Sullivan, Jessica; Waxman, Sandra; Yurovsky, Daniel (9 March 2017). "A Collaborative Approach to Infant Research: Promoting Reproducibility, Best Practices, and Theory‐Building" (PDF). Infancy. 22 (4): 421–435. doi:10.1111/infa.12182. hdl:10026.1/9942. Retrieved 19 December 2018.
  16. Tyson, Charlie (14 August 2014). "Failure to Replicate". Inside Higher Ed. Retrieved 19 December 2018.
  17. Makel, Matthew C.; Plucker, Jonathan A. (1 August 2014). "Facts Are More Important Than Novelty: Replication in the Education Sciences". Educational Researcher. 43 (6): 304–316. doi:10.3102/0013189X14545513. Retrieved 19 December 2018.
  18. Stanley, T. D.; Carter, Evan C.; Doucouliagos, Hristos (2018). "What meta-analyses reveal about the replicability of psychological research". Psychological Bulletin. 144 (12): 1325–1346. doi:10.1037/bul0000169. ISSN 1939-1455. PMID 30321017.
  19. 19.0 19.1 19.2 John, Leslie K.; Loewenstein, George; Prelec, Drazen (2012-05-01). "Measuring the Prevalence of Questionable Research Practices With Incentives for Truth Telling" (PDF). Psychological Science. 23 (5): 524–532. doi:10.1177/0956797611430953. ISSN 0956-7976. PMID 22508865.
  20. "The Nine Circles of Scientific Hell". Perspectives on Psychological Science. 7 (6): 643–644. 2012-11-01. doi:10.1177/1745691612459519. ISSN 1745-6916. PMID 26168124.
  21. Research misconduct - The grey area of Questionable Research Practices. Retrieved on 2015-11-13.
  22. Fiedler, Klaus; Schwarz, Norbert (2015-10-19). "Questionable Research Practices Revisited". Social Psychological and Personality Science. 7: 45–52. doi:10.1177/1948550615612150. ISSN 1948-5506.
  23. Simmons, Joseph; Nelson, Leif; Simonsohn, Uri (November 2011). "False-Positive Psychology: Undisclosed Flexibility in Data Collection and Analysis Allows Presenting Anything as Significant". Psychological Science. 22 (11): 1359–1366. doi:10.1177/0956797611417632. ISSN 0956-7976. PMID 22006061.
  24. Shea, Christopher (13 November 2011). "Fraud Scandal Fuels Debate Over Practices of Social Psychology". The Chronicle of Higher Education.
  25. Kahneman, Daniel. A New Etiquette for Replication.
  26. "Link to issue". Social Psychology. Hogrefe Publishing. 45 (3). 2014. Archived from the original on 30 May 2014.
  27. Table of Contents
  28. Open Science Collaboration (2015). "Estimating the reproducibility of Psychological Science" (PDF). Science. 349 (6251): aac4716. doi:10.1126/science.aac4716. hdl:10722/230596. PMID 26315443.
  29. Coyne, James (April 15, 2014). Are meta analyses conducted by professional organizations more trustworthy?. Mind the Brain. Retrieved on September 13, 2016.
  30. Connor, Steve (27 August 2015). "Study reveals that a lot of psychology research really is just 'psycho-babble'". The Independent. London.
  31. Why Psychologists' Food Fight Matters (31 July 2014).
  32. Psychology Is Starting To Deal With Its Replication Problem (27 August 2015).
  33. Science Isn't Broken (19 August 2015).
  34. Etchells, Pete (28 May 2014). "Psychology's replication drive: it's not about you". The Guardian.
  35. Wagenmakers, Eric-Jan; Wetzels, Ruud; Borsboom, Denny; Maas, Han L. J. van der; Kievit, Rogier A. (2012-11-01). "An Agenda for Purely Confirmatory Research". Perspectives on Psychological Science. 7 (6): 632–638. doi:10.1177/1745691612463078. ISSN 1745-6916. PMID 26168122.
  36. Ioannidis, John P. A. (2012-11-01). "Why Science Is Not Necessarily Self-Correcting". Perspectives on Psychological Science. 7 (6): 645–654. doi:10.1177/1745691612464056. ISSN 1745-6916. PMID 26168125.
  37. Pashler, Harold; Harris, Christine R. (2012-11-01). "Is the Replicability Crisis Overblown? Three Arguments Examined". Perspectives on Psychological Science. 7 (6): 531–536. doi:10.1177/1745691612463401. ISSN 1745-6916. PMID 26168109.
  38. Bartlett, Tom (30 January 2013). "Power of Suggestion". The Chronicle of Higher Education.
  39. 39.0 39.1 Chambers, Chris (10 June 2014). "Physics envy: Do 'hard' sciences hold the solution to the replication crisis in psychology?". The Guardian.
  40. 40.0 40.1 Makel, Matthew C.; Plucker, Jonathan A.; Hegarty, Boyd (2012-11-01). "Replications in Psychology Research How Often Do They Really Occur?". Perspectives on Psychological Science. 7 (6): 537–542. doi:10.1177/1745691612460688. ISSN 1745-6916. PMID 26168110.
  41. Collaboration, Open Science (2015-08-28). "Estimating the reproducibility of psychological science" (PDF). Science. 349 (6251): aac4716. doi:10.1126/science.aac4716. hdl:10722/230596. ISSN 0036-8075. PMID 26315443.
  42. "The Science Behind Social Science Gets Shaken Up—Again". WIRED. Retrieved 2018-08-28.
  43. Camerer, Colin F.; Dreber, Anna; et al. (27 August 2018). "Evaluating the replicability of social science experiments in Nature and Science between 2010 and 2015" (PDF). Nature Human Behaviour. 2 (9): 637–644. doi:10.1038/s41562-018-0399-z. PMID 31346273.
  44. Klein, R.A. (2018). "Many Labs 2: Investigating Variation in Replicability Across Samples and Settings". Advances in Methods and Practices in Psychological Science. 1 (4): 443–490. doi:10.1177/2515245918810225.
  45. Witkowski, Tomasz (2019). "Is the glass half empty or half full? Latest results in the replication crisis in Psychology". Skeptical Inquirer. 43 (2): 5–6.
  46. See also Earp and Trafimow, 2015
  47. 47.0 47.1 Everett, Jim Albert Charlton; Earp, Brian D. (2015-01-01). "A tragedy of the (academic) commons: interpreting the replication crisis in psychology as a social dilemma for early-career researchers". Frontiers in Psychology. 6: 1152. doi:10.3389/fpsyg.2015.01152. PMC 4527093. PMID 26300832.
  48. Resolving the replication crisis in social psychology? A new proposal. Society for Personality and Social Psychology. Retrieved on 2015-11-18.
  49. Believe it or not: how much can we rely on published data on potential drug targets? by Florian Prinz, Thomas Schlange & Khusru Asadullah, in: Nature Reviews Drug Discovery volume10, page712 (2011)
  50. Ioannidis JA (13 July 2005). "Contradicted and initially stronger effects in highly cited clinical research". JAMA. 294 (2): 218–228. doi:10.1001/jama.294.2.218. PMID 16014596.
  51. Glick, J. Leslie (1 January 1992). "Scientific data audit—A key management tool". Accountability in Research. 2 (3): 153–168. doi:10.1080/08989629208573811.
  52. Begley, C. G.; Ellis, L. M. (2012). "Drug Development: Raise Standards for Preclinical Cancer Research". Nature. 483 (7391): 531–533. Bibcode:2012Natur.483..531B. doi:10.1038/483531a. PMID 22460880.
  53. Begley, C. G. (2013). "Reproducibility: Six red flags for suspect work". Nature. 497 (7450): 433–434. Bibcode:2013Natur.497..433B. doi:10.1038/497433a.
  54. Ioannidis, JPA (2016). "Why Most Clinical Research Is Not Useful". PLoS Med. 13 (6): e1002049. doi:10.1371/journal.pmed.1002049. PMC 4915619. PMID 27328301.
  55. Ioannidis, John P. A. (August 1, 2005). "Why Most Published Research Findings Are False". PLoS Medicine. 2 (8): e124. doi:10.1371/journal.pmed.0020124. ISSN 1549-1277. PMC 1182327. PMID 16060722.
  56. Hunter, John E. (2001-06-01). "The desperate need for replications". Journal of Consumer Research. 28 (1): 149–158. doi:10.1086/321953.
  57. Armstrong, J. S.; Green, Kesten (30 January 2017). "Guidelines for Science: Evidence and Checklists". Working Paper.
  58. Aichner, Thomas; Coletti, Paolo; Forza, Cipriano; Perkmann, Urban; Trentin, Alessio (2016-03-22). "Effects of Subcultural Differences on Country and Product Evaluations: A Replication Study". Journal of Global Marketing. 29 (3): 115–127. doi:10.1080/08911762.2015.1138012.
  59. Camerer, Colin F.; Dreber, Anna; Forsell, Eskil; Ho, Teck-Hua; Huber, Jürgen; Johannesson, Magnus; Kirchler, Michael; Almenberg, Johan; Altmejd, Adam (2016-03-25). "Evaluating replicability of laboratory experiments in economics". Science. 351 (6280): 1433–1436. Bibcode:2016Sci...351.1433C. doi:10.1126/science.aaf0918. ISSN 0036-8075. PMID 26940865.
  60. "About 40% of economics experiments fail replication survey". Science. 2016-03-03. Retrieved 2017-10-25.
  61. Ioannidis, John P. A.; Stanley, T. D.; Doucouliagos, Hristos (2017-10-01). "The Power of Bias in Economics Research". The Economic Journal. 127 (605): F236–F265. doi:10.1111/ecoj.12461. ISSN 1468-0297.
  62. Halperin, Israel; Vigotsky, Andrew D.; Foster, Carl; Pyne, David B. (2018-02-01). "Strengthening the Practice of Exercise and Sport-Science Research". International Journal of Sports Physiology and Performance. 13 (2): 127–134. doi:10.1123/ijspp.2017-0322. ISSN 1555-0273. PMID 28787228.
  63. "How Shoddy Statistics Found A Home In Sports Research". FiveThirtyEight. 2018-05-16. Retrieved 2018-05-16.
  64. Stagge, James H.; Rosenberg, David E.; Abdallah, Adel M.; Akbar, Hadia; Attallah, Nour A.; James, Ryan (2019-02-26). "Assessing data availability and research reproducibility in hydrology and water resources". Scientific Data. 6: 190030. Bibcode:2019NatSD...690030S. doi:10.1038/sdata.2019.30. ISSN 2052-4463. PMC 6390703. PMID 30806638.
  65. https://www.nature.com/articles/sdata201930#f2
  66. https://replicationnetwork.com/2019/03/01/surveying-reproducibility/
  67. [ https://fabiusmaximus.com/2016/04/19/replication-crisis-in-science-95394/ The replication crisis in science has just begun. It will be big.] by Larry Kummer, Editor Science & Nature 19 April 2016 on Fabius Maximus website, here referring to W.Wilson – Scientific Regress (see Further readings)
  68. Ulf Kleinevoß: Bestimmung der Newtonschen Gravitationskonstanten, Dissertation Januar 2002, Wuppertal, S.1, Abstract; [1]
  69. Rupert Sheldrake: The Science Delusion – Freeing the Spirit of Enquiry; London 2012, Hodder& Stoughton, ISBN 978 1 444 72795 1. Chapter 3: Are the Laws of Nature Fixed?
  70. [2] and Sheldrake, Science Delusion
  71. Peter Boghossian, Ed.D. (aka Peter Boyle, Ed.D.), James Lindsay, Ph.D. (aka, Jamie Lindsay, Ph.D.): The Conceptual Penis as a Social Construct: A Sokal-Style Hoax on Gender Studies. SKEPTIC, 19.05.2017.[3]; Alexander Durin: Fehler im System mancher Wissenschaften. Telepolis, Heise, 02.03.2014. [4]; Alan D. Sokal: Transgressing the Boundaries: Towards a Transformative Hermeneutics of Quantum Gravity. Social Text 46/47:217-252, 1996. [5].
  72. Richard Horton – Offline: What is medicine’s 5 sigma? In: The Lancet VOLUME 385, ISSUE 9976, P1380, APRIL 11, 2015; [6]
  73. De Solla Price; Derek J. (1963). Little science big science. Columbia University Press.
  74. Siebert, S.; Machesky, L. M. & Insall, R. H. (2015). "Overflow in science and its implications for trust". eLife. 4: e10825. doi:10.7554/eLife.10825. PMC 4563216. PMID 26365552.
  75. Della Briotta Parolo, P.; Kumar Pan; R. Ghosh; R. Huberman; B.A. Kimmo Kaski; Fortunato, S. (2015). "Attention decay in science". Journal of Informetrics. 9 (4): 734–745. arXiv:1503.01881. Bibcode:2015arXiv150301881D. doi:10.1016/j.joi.2015.07.006.
  76. Mirowski, P. (2011). Science-Mart: Privatizing American Science. Harvard University Press.
  77. Saltelli, A.; Funtowicz, S. (2017). "What is science's crisis really about?". Futures. 91: 5–11. doi:10.1016/j.futures.2017.05.010.
  78. Benessia, A.; Funtowicz, S.; Giampietro, M.; Guimarães Pereira, A.; Ravetz, J.; Saltelli, A.; Strand, R.; van der Sluijs, J. (2016). The Rightful Place of Science: Science on the Verge. Consortium for Science, Policy and Outcomes at Arizona State University.
  79. Saltelli, Andrea; Ravetz, Jerome R. & Funtowicz, Silvio (25 June 2016). "A new community for science". New Scientist. No. 3079. p. 52.
  80. 80.0 80.1 Andrea, Saltelli (December 2018). "Why science's crisis should not become a political battling ground". Futures. 104: 85–90. doi:10.1016/j.futures.2018.07.006.
  81. Academic signaling and the post-truth world. Stony Brook University (2016-12-14). Retrieved on 5 November 2017.
  82. John Bohannon: Who’s Afraid of Peer Review?, Science 04 Oct 2013: Vol. 342, Issue 6154, pp. 60-65; DOI: 10.1126/science.342.6154.60
  83. Nissen, Silas Boye; Magidson, Tali; Gross, Kevin; Bergstrom, Carl (December 20, 2016). "Research: Publication bias and the canonization of false facts". eLife. 5: e21451. arXiv:1609.00494. doi:10.7554/eLife.21451. Retrieved 9 June 2019.
  84. Rupert Sheldrake: How the Universal Gravitational Constant Varies; The Replicability Crisis in Science; Science Set Free: 10 Paths to New Discovery, ISBN-13: 978-0770436704, Publ. Deepak Chopra; 1st ed. (4. September 2012); The Science Delusion, ISBN-13: 978-1444727944, Publisher: Coronet (6 Dec. 2012)
  85. Oreskes, N. (2018). "Beware: Transparency rule is a trojan horse". Nature. 557 (7706): 469. Bibcode:2018Natur.557..469O. doi:10.1038/d41586-018-05207-9. PMID 29789751.
  86. Michaels, D. (2008). Doubt is their product: How industry's assault on science threatens your health. Oxford University Press.
  87. See all of the cited literature.
  88. Moonesinghe, Ramal; Khoury, Muin J.; Janssens, A. Cecile J. W. (2007-02-27). "Most Published Research Findings Are False—But a Little Replication Goes a Long Way". PLoS Med. 4 (2): e28. doi:10.1371/journal.pmed.0040028. PMC 1808082. PMID 17326704.
  89. Simons, Daniel J. (2014-01-01). "The Value of Direct Replication". Perspectives on Psychological Science. 9 (1): 76–80. doi:10.1177/1745691613514755. ISSN 1745-6916. PMID 26173243.
  90. Ioannidis, John P. A.; Fanelli, Daniele; Dunne, Debbie Drake; Goodman, Steven N. (2015-10-02). "Meta-research: Evaluation and Improvement of Research Methods and Practices". PLOS Biology. 13 (10): –1002264. doi:10.1371/journal.pbio.1002264. ISSN 1545-7885. PMC 4592065. PMID 26431313.
  91. On communicating science and uncertainty: A podcast with John Ioannidis (8 December 2015). Retrieved on 20 May 2019.
  92. Registered Replication Reports. Association for Psychological Science. Retrieved on 2015-11-13.
  93. Chambers, Chris (2014-05-20). "Psychology's 'registration revolution'". The Guardian. Retrieved 2015-11-13.
  94. Lindsay, D. Stephen (2015-11-09). "Replication in Psychological Science". Psychological Science. 26 (12): 1827–32. doi:10.1177/0956797615616374. ISSN 0956-7976. PMID 26553013.
  95. Yeung, Andy W. K. (2017). "Do Neuroscience Journals Accept Replications? A Survey of Literature". Frontiers in Human Neuroscience. 11: 468. doi:10.3389/fnhum.2017.00468. ISSN 1662-5161. PMC 5611708. PMID 28979201.
  96. Martin, G. N.; Clarke, Richard M. (2017). "Are Psychology Journals Anti-replication? A Snapshot of Editorial Practices". Frontiers in Psychology. 8: 523. doi:10.3389/fpsyg.2017.00523. ISSN 1664-1078. PMC 5387793. PMID 28443044.
  97. Frank, Michael C.; Saxe, Rebecca (2012-11-01). "Teaching Replication". Perspectives on Psychological Science. 7 (6): 600–604. doi:10.1177/1745691612460686. ISSN 1745-6916. PMID 26168118.
  98. Grahe, Jon E.; Reifman, Alan; Hermann, Anthony D.; Walker, Marie; Oleson, Kathryn C.; Nario-Redmond, Michelle; Wiebe, Richard P. (2012-11-01). "Harnessing the Undiscovered Resource of Student Research Projects". Perspectives on Psychological Science. 7 (6): 605–607. doi:10.1177/1745691612459057. ISSN 1745-6916. PMID 26168119.
  99. Benjamin, Daniel. "Redefine statistical significance". PsyArXiv.
  100. Colquhoun, David (2015). "An investigation of the false discovery rate and the misinterpretation of p-values". Royal Society Open Science. 1 (3): 140216. arXiv:1407.5296. Bibcode:2014RSOS....140216C. doi:10.1098/rsos.140216. PMC 4448847. PMID 26064558.
  101. 101.0 101.1 101.2 101.3 Colquhoun, David (2017). "The reproducibility of research and the misinterpretation of p-values". Royal Society Open Science. 4 (12): 171085. doi:10.1098/rsos.171085. PMC 5750014. PMID 29308247.
  102. The problem with p-values. Aeon Magazine. Retrieved on 11 December 2016.
  103. Calculator for false positive risk (FPR). UCL.
  104. Matthews, R. A. J. (2001). "Why should clinicians care about Bayesian methods?". Journal of Statistical Planning and Inference. 94: 43–58. doi:10.1016/S0378-3758(00)00232-9.
  105. Maxwell, Scott E.; Lau, Michael Y.; Howard, George S. (2015). "Is psychology suffering from a replication crisis? What does "failure to replicate" really mean?". American Psychologist. 70 (6): 487–498. doi:10.1037/a0039400. PMID 26348332.
  106. IntHout, Joanna; Ioannidis, John P. A.; Borm, George F.; Goeman, Jelle J. (2015). "Small studies are more heterogeneous than large ones: a meta-meta-analysis". Journal of Clinical Epidemiology. 68 (8): 860–869. doi:10.1016/j.jclinepi.2015.03.017. PMID 25959635.
  107. 107.0 107.1 Button, Katherine S.; Ioannidis, John P. A.; Mokrysz, Claire; Nosek, Brian A.; Flint, Jonathan; Robinson, Emma S. J.; Munafò, Marcus R. (2013-05-01). "Power failure: why small sample size undermines the reliability of neuroscience". Nature Reviews Neuroscience. 14 (5): 365–376. doi:10.1038/nrn3475. ISSN 1471-003X. PMID 23571845.
  108. Greenwald, Anthony G. (1975). "Consequences of prejudice against the null hypothesis". Psychological Bulletin. 82 (1): 1–20. doi:10.1037/h0076157.
  109. Amrhein, Valentin; Korner-Nievergelt, Fränzi; Roth, Tobias (2017). "The earth is flat (p > 0.05): significance thresholds and the crisis of unreplicable research". PeerJ. 5: e3544. doi:10.7717/peerj.3544. PMC 5502092. PMID 28698825.
  110. 110.0 110.1 Ioannidis, John P. A. (2016). "Anticipating consequences of sharing raw data and code and of awarding badges for sharing". Journal of Clinical Epidemiology. 70: 258–260. doi:10.1016/j.jclinepi.2015.04.015. PMID 26163123.
  111. NWO makes 3 million available for Replication Studies pilot. Retrieved on 2 August 2016.
  112. 112.0 112.1 112.2 Apple, Sam (January 22, 2017). "The Young Billionaire Behind the War on Bad Science". Wired.
  113. Munafò, Marcus R.; Smith, George Davey (January 23, 2018). "Robust research needs many lines of evidence". Nature. 553 (7689): 399–401. Bibcode:2018Natur.553..399M. doi:10.1038/d41586-018-01023-3. PMID 29368721.