Draft:Questionable research practices
![]() | Draft article not currently submitted for review.
This is a draft Articles for creation (AfC) submission. It is not currently pending review. While there are no deadlines, abandoned drafts may be deleted after six months. To edit the draft click on the "Edit" tab at the top of the window. To be accepted, a draft should:
It is strongly discouraged to write about yourself, your business or employer. If you do so, you must declare it. Where to get help
How to improve a draft
You can also browse Wikipedia:Featured articles and Wikipedia:Good articles to find examples of Wikipedia's best writing on topics similar to your proposed article. Improving your odds of a speedy review To improve your odds of a faster review, tag your draft with relevant WikiProject tags using the button below. This will let reviewers know a new draft has been submitted in their area of interest. For instance, if you wrote about a female astronomer, you would want to add the Biography, Astronomy, and Women scientists tags. Editor resources
Last edited by Onel5969 (talk | contribs) 8 days ago. (Update) |
![]() | This article is written like a personal reflection, personal essay, or argumentative essay that states a Wikipedia editor's personal feelings or presents an original argument about a topic. (March 2025) |
![]() | It has been suggested that this page be merged into Scientific integrity. (Discuss) Proposed since March 2025. |
Questionable research practices (QRPs) are behaviors undertaken during the design, implementation, analysis and dissemination of scientific research that are unscientific. These range from grey area behaviors that tweak or oversell findings to outright violations of scientific ethics and research integrity if not laws against data fraud. These behaviors deviate from ideal scientific conduct.[1] QRPs increase uncertainty in scientific research. Generally QRPs increase the likelihood of false positives[2] and contribute to the replication crisis across scientific fields.
Studies of QRPs thus far confirm they are widespread.[3][4] There are debates over whether "questionable" includes outright data or result faking or not,[5] but most studies of QRPs include these as the most extreme types.[3][6][7] Moreover, fraud does not have a clear cutoff point. For example, rounding a p-value in the wrong direction is fraudulent, but clearly not as severe as faking experimental data. A Google Ngram search reveals that the term QRP entered scientific usage in the late 1980s and exploded into mainstream science after 2010. The 2010 decade began with some scandals that probably contributed to this. In psychology a well known and respected professor was found to have faked an entire career[8] of research and in one of the most highly respected journals another psychologist published[9] scientific evidence of extra-sensory perception (ESP).
Definition and scope
[edit]These include any actions that skew research outcomes or interpretations by violating standard methodological or transparency norms. They are behaviors that deviate from ideal scientific practices, such as the scientific method and scientific integrity. In some circles they encompass blatant fraud, such as faking data or images.[7] Across all definitions, they primarily involve subtle practices often related to analysis and reporting that can mislead without outright lying. For example, intentionally excluding data points or cases after seeing their effect on results, or failing to report entire studies that yielded negative results
Such behaviors inflate evidence in favor of a hypothesis spuriously and undermine the integrity of findings. The term gained wider usage after a 2012 study by Leslie K. John, George Loewenstein, and Dražen Prelec, who surveyed researchers about these behaviors and found that over half of active psychologists surveyed, reported engaging in at least one QRP during their career.[3] By surveying researchers about their own practices and those of their peers, they used a Bayesian truth serum method to conclude modestly that 9% of psychologists likely faked data during their career. John et al. and others emphasized that the majority of QRPs, by virtue of being “questionable” rather than overtly fraudulent, allow considerable room for self-justification.
A researcher might convince themselves that dropping an outlier or tweaking a hypothesis post hoc is defensible, even if it biases the result. This capacity for rationalization makes QRPs insidious: many scientists may engage in them without feeling they are doing anything wrong. As a result, the prevalence of QRPs can be high even among otherwise honest, well-intentioned researchers.[citation needed]
QRPs cover a range of specific practices at different stages of the research process. They are predominantly known in research involving experiments and statistical analysis and reporting, but are equally possible in qualitative and ethnographic research albeit in different formats. The following is a general list of QRPs expanded from Peter M. Dahlgren's Chapter 2 in How Scientists Lie expanded to include fraud and qualitative questionable research practices.
P-hacking
[edit]Manipulating the research process to obtain certain results: P-hacking is among the most prominent QRPs with research involving statistical hypothesis testing with a p-value (probability value). Researchers re-run, readjust and reanalyze models until the p-value reaches a certain cutoff value. Although the name 'p-hacking' comes from the 'p' in p-value, it is used today in general to reference behaviors that manipulate the research process to achieve statistical effects that are 'significant'[10] - for example by tinkering with models or torturing original data. It could refer to intentionally searching for insignificant results or 'null hacking'.[11] What counts as a significant p-value is a product of the nature of the hypothesis, convention and/or power analysis. This behavior produces a result that is likely an outlier or unique occurrence, rather than a reliable scientific test of anything.
Selective reporting
[edit]Manipulating the process of reporting after conducting research: Selectively disclosing or emphasizing only those results that are desirable, for example statistically significant and/or pointing in the 'right' direction. It also includes dropping carefully derived hypotheses in case results do not support them, as if they never existed. This act of cherry-picking can occur at the level of outcomes (reporting only significant dependent measures and ignoring non-significant ones), analyses (reporting only analyses that “worked”), or even entire studies (the so-called file drawer problem[12] or publication bias, where experiments that yielded null or undesirable results are never reported). By omitting inconsistent data, researchers create a biased picture of support for a hypothesis.
HARKing
[edit]Manipulating the order of events in the research process: Hypothesizing after results are known (HARKing) is a method of explaining results post hoc. Either the results were something different than hypothesized or resulted from playing around with data exploratively. But in either case, the hypotheses are presented as if they were established a priori. This presents the results as confirmatory rather than exploratory. This can have grave policy implications and inflate the reliability of a given effect or theory.[13]
Publication bias
[edit]Scientific vetting of results submitted to journals tends to favor 'sexy' findings, those that will garner more attention and boost citation counts.[14][15] This leads to editors and peer reviewers to favor finds with large effect sizes, small p-values or generally surprising or shocking findings that garner attention. Bias in the process of selection of what gets published generates perverse incentives for researchers to p-hack, selectively report and HARK.[16]
Salami slicing
[edit]Carving a single study up into the most possible publishable papers. This dilutes the scientific findings pool by generating an appearance that a single hypothesis has more studies supporting it than there actually are.[17]
Citation bias
[edit]The practice of disproportionally citing studies that support an author's theory or hypothesis.[18] This generates a distorted picture of the existing scientific literature and can generate bias in what studies are cited and read.
Fraud
[edit]Any direct manipulation of results. Not in any way defensible. For example inventing entire datasets, adding cases or changing the values of existing cases are examples of this. The Stapel and LaCour cases involved entirely fabricated experiments and resulting data, whereas the Gino and Stewart retractions involved manipulation of original data. Fraud includes various formats of data-faking. Many retractions occur because images were AI generated or manipulated.[19]
Sloppy science
[edit]Often it is difficult to determine if researchers consciously manipulated the research practice, or were trained to behave in questionable ways. Sometimes researchers are simply sloppy. This appears to be the case in the Reinhart-Rogoff retraction which seems to have been a product of copy-pasting errors. Sometimes researchers rush and introduce errors into their workflows that render the results unreliable.[20]
Historical background and growth of awareness
[edit]Early concerns about what we now call QRPs can be traced back decades, but awareness grew substantially in the 2000s and 2010s. In clinical research and psychology, scientists began noticing that published findings often seemed too good to be true, and that many results could not be reproduced. In 2005, epidemiologist John Ioannidis published a provocative article titled “Why Most Published Research Findings Are False,” arguing that a combination of QRPs like selective reporting and publication bias meant that much of the published literature was unreliable.[21] Although a thought experiment, the arguments caught hold. Around the same time, surveys were documenting questionable behaviors among researchers: for example, a 2005 survey of 3,247 U.S. scientists found that one-third admitted to having engaged in at least one dubious practice in the past three years.[1] These early warnings set the stage for more systematic scrutiny of research practices. In psychology, a series of events in the early 2010s dramatically increased attention to QRPs. In 2011, Joseph Simmons, Leif Nelson, and Uri Simonsohn published a landmark paper titled “False-Positive Psychology,” which demonstrated through simulations and actual experiments that common flexible analytic practices can lead to spurious significance. When coupled with the aforementioned Bem and Stapel scandals, psychology was at a major turning point.[22]
The awareness of QRPs spread across the social and behavioral sciences. For example, in Political Science, the American Journal of Political Science began testing all code for computational reproducibility in 2015 for accepted articles using statistical analysis. None of it reproduced at first. This coupled with the aforementioned LaCour scandal had a deep impact in this discipline.[23] In Economics, work by Abel Brodeur and colleagues made painfully clear that p-hacking and publication biases were undermining the reliability of the discipline.[24][25][26]
By the mid-2010s, concerns about QRPs had crystallized into what is now called the replication crisis. Systematic replication projects were undertaken to assess how many findings would hold up if re-tested independently. The results were sobering: in 2015, the Open Science Collaboration repeated 100 published psychology experiments and found that only ~36% produced a statistically significant result again, and the effect sizes were on average about half of the originals.[27] Similar alarm bells rang in other fields – for instance, in cancer biology, pharmaceutical companies reported they could not replicate many high-profile preclinical studies. These fields were grappling with this in serious ways because their findings literally shaped life and death for humans depending on medical treatments.
A landmark survey by John, Loewenstein, and Prelec (2012) shed light on QRPs within Psychology, although there is little reason to believe other disciplines that work with experiments and/or statistical analysis are so different.[28] Some specific behaviors were strikingly common: for example, around 66% of surveyed psychologists acknowledged selectively not reporting all of a study’s dependent measures (variables) in publication, and roughly 50% admitted to only reporting studies that gave the desired result (“cherry-picking” successful experiments). Using Bayesian truth serum methods they estimated that 94% of researchers engaged in at least one QRP during their career.
Metascience (the scientific study of science itself) and replication research exploded in scope to study these problems.[29] By the late 2010s, terms like “p-hacking” and “HARKing” had entered mainstream discussions of research reliability, and journals and funding agencies began exploring ways to curb such practices. In summary, what began as scattered concerns in the early 2000s evolved into a full-blown crisis of confidence in research findings, with questionable research practices identified as a key underlying cause. This realization has driven a reform movement aiming to improve transparency and reproducibility in science. Although there are debates over the severity of the 'crisis' the metascience is clear that science is not as reliable, and scientists sometimes not as trustworthy as the public, government officials and other scientists like to think.[30]
Causes and incentives for QRPs
[edit]QRPs are often driven by the incentive structure of academia and science – in other words, the pressures and rewards that researchers face. An academic scientist’s career success depends heavily on publishing research, particularly in high-impact journals, which in turn favor novel and statistically significant findings.[31] This creates a strong motivation to obtain positive results and avoid null findings. Experiments or analyses that “don’t work” (i.e. produce non-significant results) are less likely to be published due to publication bias, leading researchers to perceive such outcomes as failures. The adage “publish or perish” encapsulates this environment: to secure jobs, promotions, and grants, scientists feel pressure to publish frequently and in prestigious venues, which typically means showing exciting, confirmatory results. As Nosek and colleagues noted in 2012,[31] disciplinary incentives encourage decisions that produce positive outcomes and gloss over negative or inconclusive results.
Gary King pointed out in his 'How Not to Lie With Statistics' paper that "often, we learn each others' mistakes rather than learning from each others' mistakes.".[32] We cannot entirely blame researchers for simply doing things the way they are taught. This means that institutions, supervisors, curricula and publishing outlets play active roles in promoting scientific norms favoring QRPs.
References
[edit]- ^ a b Brooker, Robin; Allum, Nick (2024-10-14). "Investigating the links between questionable research practices, scientific norms and organisational culture". Research Integrity and Peer Review. 9 (1): 12. doi:10.1186/s41073-024-00151-x. ISSN 2058-8615. PMC 11472529. PMID 39397013.
- ^ Simmons, Joseph P.; Nelson, Leif D.; Simonsohn, Uri (17 October 2011). "False-Positive Psychology: Undisclosed Flexibility in Data Collection and Analysis Allows Presenting Anything as Significant". Psychological Science. 22 (11): 1359–1366. doi:10.1177/0956797611417632. ISSN 0956-7976. PMID 22006061.
- ^ a b c John, Leslie K.; Loewenstein, George; Prelec, Drazen (May 2012). "Measuring the Prevalence of Questionable Research Practices With Incentives for Truth Telling". Psychological Science. 23 (5): 524–532. doi:10.1177/0956797611430953. ISSN 0956-7976. PMID 22508865.
- ^ Gopalakrishna, Gowri; Riet, Gerben ter; Vink, Gerko; Stoop, Ineke; Wicherts, Jelte M.; Bouter, Lex M. (2022-02-16). "Prevalence of questionable research practices, research misconduct and their potential explanatory factors: A survey among academic researchers in The Netherlands". PLOS ONE. 17 (2): e0263023. Bibcode:2022PLoSO..1763023G. doi:10.1371/journal.pone.0263023. ISSN 1932-6203. PMC 8849616. PMID 35171921.
- ^ Resnik, David B. (2023), Valdés, Erick; Lecaros, Juan Alberto (eds.), "Research Misconduct and Questionable Research Practices", Handbook of Bioethical Decisions. Volume II: Scientific Integrity and Institutional Ethics, Collaborative Bioethics, vol. 3, Cham: Springer International Publishing, pp. 25–36, doi:10.1007/978-3-031-29455-6_3, ISBN 978-3-031-29455-6, retrieved 2025-03-30
- ^ Gopalakrishna, Gowri; ter Riet, Gerben; Vink, Gerko; Stoop, Ineke; Wicherts, Jelte M.; Bouter, Lex M. (2022-02-16). Fàbregues, Sergi (ed.). "Prevalence of questionable research practices, research misconduct and their potential explanatory factors: A survey among academic researchers in The Netherlands". PLOS ONE. 17 (2): e0263023. Bibcode:2022PLoSO..1763023G. doi:10.1371/journal.pone.0263023. ISSN 1932-6203. PMC 8849616. PMID 35171921.
- ^ a b Fanelli, Daniele (2009-05-29). Tregenza, Tom (ed.). "How Many Scientists Fabricate and Falsify Research? A Systematic Review and Meta-Analysis of Survey Data". PLOS ONE. 4 (5): e5738. Bibcode:2009PLoSO...4.5738F. doi:10.1371/journal.pone.0005738. ISSN 1932-6203. PMC 2685008. PMID 19478950.
- ^ Larregue, Julien (May 2024). "Sentencing social psychology: Scientific deviance and the diffusion of statistical rules". Current Sociology. 72 (3): 389–406. doi:10.1177/00113921221117604. ISSN 0011-3921.
- ^ Bem, Daryl J. (2011). "Feeling the future: Experimental evidence for anomalous retroactive influences on cognition and affect". Journal of Personality and Social Psychology. 100 (3): 407–425. doi:10.1037/a0021524. ISSN 1939-1315. PMID 21280961.
- ^ Stefan, Angelika M.; Schönbrodt, Felix D. (2023-02-08). "Big little lies: a compendium and simulation of p-hacking strategies". Royal Society Open Science. 10 (2): 220346. Bibcode:2023RSOS...1020346S. doi:10.1098/rsos.220346. PMC 9905987. PMID 36778954.
- ^ Chuard, Pierre J. C.; Vrtílek, Milan; Head, Megan L.; Jennions, Michael D. (2019-01-25). "Evidence that nonsignificant results are sometimes preferred: Reverse P-hacking or selective reporting?". PLOS Biology. 17 (1): e3000127. doi:10.1371/journal.pbio.3000127. ISSN 1545-7885. PMC 6364929. PMID 30682013.
- ^ Moniz, Philip; Druckman, James N.; Freese, Jeremy (2025-03-25). "The file drawer problem in social science survey experiments". Proceedings of the National Academy of Sciences. 122 (12): e2426937122. doi:10.1073/pnas.2426937122. ISSN 0027-8424. PMC 11962440. PMID 40106354.
- ^ Rubin, Mark (2017-12-01). "When Does HARKing Hurt? Identifying When Different Types of Undisclosed Post Hoc Hypothesizing Harm Scientific Progress". Review of General Psychology. 21 (4): 308–320. doi:10.1037/gpr0000128. ISSN 1089-2680.
- ^ Ioannidis, John P.A.; Munafò, Marcus R.; Fusar-Poli, Paolo; Nosek, Brian A.; David, Sean P. (May 2014). "Publication and other reporting biases in cognitive sciences: detection, prevalence, and prevention". Trends in Cognitive Sciences. 18 (5): 235–241. doi:10.1016/j.tics.2014.02.010. PMC 4078993. PMID 24656991.
- ^ Gerber, Alan S.; Malhotra, Neil (2008-08-01). "Publication Bias in Empirical Sociological Research: Do Arbitrary Significance Levels Distort Published Results?". Sociological Methods & Research. 37 (1): 3–30. doi:10.1177/0049124108318973. ISSN 0049-1241.
- ^ Blanco-Perez, Cristina; Brodeur, Abel (2020-07-01). "Publication Bias and Editorial Statement on Negative Findings". The Economic Journal. 130 (629): 1226–1247. doi:10.1093/ej/ueaa011. ISSN 0013-0133.
- ^ Adams, Nicholas Norman (2022-06-16). "Salami Slicing: clarifying common misconceptions for social science early-career researchers". SN Social Sciences. 2 (7): 88. doi:10.1007/s43545-022-00389-6. ISSN 2662-9283.
- ^ Gøtzsche, Peter C (2022-01-01). "Citation bias: questionable research practice or scientific misconduct?". Journal of the Royal Society of Medicine. 115 (1): 31–35. doi:10.1177/01410768221075881. ISSN 0141-0768. PMC 8814996. PMID 35105192.
- ^ Gu, Jinjin; Wang, Xinlei; Li, Chenang; Zhao, Junhua; Fu, Weijin; Liang, Gaoqi; Qiu, Jing (2022-07-08). "AI-enabled image fraud in scientific publications". Patterns (New York, N.Y.). 3 (7): 100511. doi:10.1016/j.patter.2022.100511. ISSN 2666-3899. PMC 9278510. PMID 35845832.
- ^ Breznau, Nate; Rinke, Eike Mark; Wuttke, Alexander; Adem, Muna; Adriaans, Jule; Akdeniz, Esra; Alvarez-Benjumea, Amalia; Andersen, Henrik K.; Auer, Daniel; Azevedo, Flavio; Bahnsen, Oke; Bai, Ling; Balzer, Dave; Bauer, Paul C.; Bauer, Gerrit (March 2025). "The reliability of replications: a study in computational reproductions". Royal Society Open Science. 12 (3). Bibcode:2025RSOS...1241038B. doi:10.1098/rsos.241038. ISSN 2054-5703. PMC 11922520. PMID 40109936.
- ^ Ioannidis, John P. A. (2005-08-30). "Why Most Published Research Findings Are False". PLOS Medicine. 2 (8): e124. doi:10.1371/journal.pmed.0020124. ISSN 1549-1676. PMC 1182327. PMID 16060722.
- ^ Larsen, Sally A. (2021-04-22). "Book Review: Science Fictions: Exposing Fraud, Bias, Negligence and Hype in Science". Frontiers in Psychology. 12. doi:10.3389/fpsyg.2021.666441. ISSN 1664-1078.
- ^ politicalsciencereplication.wordpress.com (2015-05-04). "Leading journal verifies articles before publication – so far, all replications failed". Political Science Replication. Retrieved 2025-03-30.
- ^ Brodeur, Abel; Carrell, Scott; Figlio, David; Lusher, Lester (2023-11-01). "Unpacking p-Hacking and Publication Bias". American Economic Review. 113 (11): 2974–3002. doi:10.1257/aer.20210795. hdl:10419/274141. ISSN 0002-8282.
- ^ Brodeur, Abel; Cook, Nikolai; Neisser, Carina (2024-04-01). "p-Hacking, Data type and Data-Sharing Policy". The Economic Journal. 134 (659): 985–1018. doi:10.1093/ej/uead104. ISSN 0013-0133.
- ^ Brodeur, Abel; Cook, Nikolai; Heyes, Anthony (2020-11-01). "Methods Matter: p-Hacking and Publication Bias in Causal Analysis in Economics". American Economic Review. 110 (11): 3634–3660. doi:10.1257/aer.20190687. ISSN 0002-8282.
- ^ Open Science Collaboration (2015-08-28). "Estimating the reproducibility of psychological science". Science. 349 (6251): aac4716. doi:10.1126/science.aac4716. hdl:10722/230596. PMID 26315443.
- ^ John, Leslie K.; Loewenstein, George; Prelec, Drazen (2012-05-01). "Measuring the Prevalence of Questionable Research Practices With Incentives for Truth Telling". Psychological Science. 23 (5): 524–532. doi:10.1177/0956797611430953. ISSN 0956-7976. PMID 22508865.
- ^ Edlund, John E.; Cuccolo, Kelly; Irgens, Megan S.; Wagge, Jordan R.; Zlokovich, Martha S. (January 2022). "Saving Science Through Replication Studies". Perspectives on Psychological Science. 17 (1): 216–225. doi:10.1177/1745691620984385. ISSN 1745-6916. PMID 33682522.
- ^ Breznau, Nate (2021-01-25). "Does Sociology Need Open Science?". Societies. 11 (1): 9. doi:10.3390/soc11010009. ISSN 2075-4698.
- ^ a b Nosek, Brian A.; Spies, Jeffrey R.; Motyl, Matt (November 2012). "Scientific Utopia: II. Restructuring Incentives and Practices to Promote Truth Over Publishability". Perspectives on Psychological Science. 7 (6): 615–631. doi:10.1177/1745691612459058. ISSN 1745-6916. PMC 10540222. PMID 26168121.
- ^ King, Gary (1986). "How Not to Lie with Statistics: Avoiding Common Mistakes in Quantitative Political Science". American Journal of Political Science. 30 (3): 666–687. doi:10.2307/2111095. ISSN 0092-5853. JSTOR 2111095.