Psychology’s Meta-Analysis Problem
Psychology has a meta-analysis problem. And that’s contributing to its reproducibility problem. Meta-analyses are wallpapering over many research weaknesses, instead of being used to systematically pinpoint them.
It’s a bit of a case of a prophet not being recognized in its hometown. Meta-analysis was born in the late 1970s in educational psychology. It’s the process of combining and analyzing data from more than one study at a time (more here).
Pretty soon, medical researchers realized it could be “a key element in improving individual research efforts and their reporting”. They forked off down the path of embedding meta-analysis in systematic reviewing – so much so, that many people mistake “meta-analysis” and “systematic review” as synonyms.
But they’re not. Systematic reviewing doesn’t just gather and crunch numbers from multiple studies. The process should include systematic critical assessment of the quality of study design and risk of bias in the data as well. Unreliable study results can fatally tilt results in a meta-analysis in misleading directions.
And psychology research can have a very high risk of bias. For example, many areas of psychology research are particularly prone to bias because of their populations: often volunteers rather than a sample selected to reduce bias, college students, or drawn from the internet or Mechanical Turk. They tend to be, as Joseph Henrich and colleagues pointed out [PDF], “the weirdest people in the world”: from Western, Educated, Industrialized, Rich, and Democratic (WEIRD) societies. And their responses may not be reliable enough, either.
Here’s an example of what this means for a body of studies. I looked at this meta-analysis on changing implicit bias when l was researching for a recent blog post. It sounds like a solid base from which to draw conclusions: 427 studies with 63,478 participants.
But 84% of the participants are college students, 66% female, and 55% from the USA. They’re nearly all very short-term studies (95%), 85% had no measures of behavior in them, and 87% had no pre-test results.
James Coyne recently walked through a range of other critical biases and “questionable research practices” (QRPs) affecting the literature in psychology, including selective reporting of analyses, studies that are too small and lack scientific rigor, over-reliance on statistical significance testing, and investigator allegiance to a therapy or idea. John Ioannidis and colleagues (2014) also document extensive over-reporting of positive results and other reporting biases across the cognitive sciences.
Coyne and colleagues found that these problems were bleeding through into a group of meta-analyses they studied, too [PDF]. Christopher Ferguson and Michael Brannick suspected a skew from unpublished studies in 20-40% of the psychology meta-analyses they studied [PDF]. Michiel van Elk and his colleagues found similar problems, pointing out
[I]f there is a true effect, then a meta-analysis will pick it up. But a meta-analysis will also indicate the existence of an effect if there is no true effect but only experimenter bias and QRPs.
Guy Cafri, Jeffrey Kromrey and Michael Brannick add to this list a range of problems of statistical rigor in meta-analyses published across a decade in a major psychology journal – including multiple testing (an explainer here).
It’s not that these problems don’t exist in meta-analysis in other fields. But the problem seems far more severe in psychology, largely, it seems to me, because of the way meta-analysis has evolved in this field. In clinical research, systematic reviewing started to make these problems clear in the 1970s and 1980s – and the community started chipping away at them.
Formal instruments for assessing the robustness and risk of bias of different study types were developed and evaluated, and those assessments were incorporated into systematic reviews. We called what drove this process methodology or methods research. John Ioannidis and his team at METRICS call it meta-research:
Meta-research is an evolving scientific discipline that aims to evaluate and improve research practices.
It’s a long way from being a perfect science itself. It’s not that easy to reliably measure the reliability of data! And it’s not implemented as fully as it should be in clinical systematic reviews either. But it is the expected standard. (See for example PRISMA and AMSTAR for systematic reviews, and Julian Higgins and colleagues for meta-analyses [PDF].) Two of the key features of systematic reviews are appraising the quality of the evidence and minimizing bias within the review process itself.
Yet, in the equivalent for psychology – Meta-Analysis Reporting Standards (MARS) – assessment of data quality is optional. Gulp.
Fortunately, in psychological therapies, many adhere to clinical research standards for systematic reviews. To test my impressions of how non-therapeutic, non-imaging meta-analyses in psychology stack up, I assessed 10 published in the psychological literature in June on some key criteria. (Details of how I got to this 10 are at the foot of this post. And for comparison with biomedical systematic reviews, I used the recent study by Matthew Page and colleagues.)
It’s a small group of psychology meta-analyses, and I’m the only person who did the assessment, which are 2 big sources of bias. I tried to be very generous. It still isn’t a pretty picture – only 1 out of 10 meta-analyses came out with more thumbs up than down (“J”). And this was only a subset of quality criteria for a systematic review.
None reported a protocol or ‘a priori’ design for the meta-analysis. That’s explicitly reported in about 40% of systematic reviews in biomedical research. The rate of comprehensive search for studies, with detail reported, was 60% (compared with around 90% for biomedicine). And the key issues of quality assessment? Half as often for psychology: only 30 and 40%, compared with 60 and 80% in biomedicine.
The risk of bias in the studies meta-analyzed was mostly unquestioned. Perhaps more of the authors share the misassumption of authors of “A”:
48.9% of the studies in this dataset (i.e. 68 out of 139 published articles) come from the six leading SCM journals mentioned above, and 77.7% of the papers (i.e. 108 out of 139) are published in journals included in the master list of the Social Sciences Citation Index. Therefore, the quality of the studies that provide the correlations for this meta-analysis is considered high. A list of the studies is available on request.
Oh boy. Perhaps people would have less faith in the published literature in their discipline if they really looked under the hood.
When I was looking for research tools to evaluate risk of bias in psychology studies, I found an article with the promising title, “Evaluating psychological reports: dimensions, reliability, and correlates of quality judgments”. But this 1978 article turned out to be instructive for other reasons. It was based on what editors and consultants for 9 major psychology journals thought was worth publishing. It rather meticulously describes the foundation on which psychology’s replication crisis was built:
Component VI seems primarily to reflect scientific advancement, while Component VII seems to merit the label “data grinders” or “brute empiricism,” with emphasis on description rather than explanation. The eighth component might be labeled “routine” or “ho-hum” research and brings to mind the remark one respondent penciled in the questionnaire margin: “How about an item like, ‘Oh God, there goes old so-and-so again!’ ?”
The derided components VII and VIII include bedrocks of good science, like “The author uses precisely the same procedures as everyone else”, and “The results are not overwhelming in their implications, but they constitute a needed component of knowledge”. You can see what this publication culture led to in a great example Neuroskeptic explained: romantic priming. Or another case – including problematic meta-analysis – Daniel Engber wrote about: self-control and ego depletion.
Meta-research is well on its way in psychology and neuroscience, and pointing to lots of ways forward. Kate Button and colleagues show the importance of increasing sample sizes, Chris Chambers and more than 80 colleagues call for study pre-registration.
Changing the expectations of meta-analysis in psychology needs to be part of the new direction, too. Meta-analysis began in psychology. But long experience is no guarantee of being on the right track. Oscar Wilde summed that up in 1890 in The Picture of Dorian Gray:
He began to wonder whether we could ever make psychology so absolute a science that each little spring of life would be revealed to us. As it was, we always misunderstood ourselves and rarely understood others. Experience was of no ethical value. It was merely the name men gave to their mistakes…All that it really demonstrated was that our future would be the same as our past.
~~~~
More on meta-analysis:
5 Tips for Understanding Data in Meta-Analysis
5 Key Things to Know About Meta-Analysis
Another 5 Things to Know About Meta-Analysis
The Forest Plot Trilogy: A Gripping Thriller Concludes
Disclosures: My ongoing PhD research is meta-research in systematic reviewing in health care, and assessing the quality of systematic reviews is part of my role at PubMed Health. I was one of the founding members of the Cochrane Collaboration and Germany’s Institute for Quality and Efficiency in Health Care (IQWiG), participated in a standard-setting group for reporting abstracts of systematic reviews, and was a member of the GRADE Working Group from 2008 to 2011.
The cartoons (including the mock reproducibility meta-analysis at the top of the post) are my own (CC-NC-ND-SA license). (More cartoons at Statistically Funny and on Tumblr.)
List of 10 meta-analyses:
A. Chang (2016)
B. Elvik (2016)
C. Hagger (2016)
D. Kabins (2016)
E. Knežević (2016)
F. Koh (2016)
G. McDermott (2016)
H. Nahman-Averbuch (2016)
I. Pouwels (2016)
J. Shoji (2016)
These were the first 10 which were not systematic reviews, therapy, gene association, survey, or imaging meta-analyses, from a PsycINFO search on 10 June 2016: meta-analysis in title. I also had to have immediate access to the full text. The full list of 37 citations it took to yield these 10 is here.
Rough comparison of outcomes between this set and Page et al’s large set of systematic reviews in biomedicine:
* The thoughts Hilda Bastian expresses here at Absolutely Maybe are personal, and do not necessarily reflect the views of the National Institutes of Health or the U.S. Department of Health and Human Services.
Nicely thought provoking and sensibly argued. I’m not really competent to assess the fine grained detail but i found this well worth a read and well worth thinking on.
Hilda:
I serendipitously found this nice post of yours and was pleasantly surprised when I clicked on the link in the phrase “Pretty soon, medical researchers realized” and found my old paper.
Now, I don’t believe it was soon or widely recognized even sometime afterward. I wrote the paper as I was very upset that my colleagues at the University of Toronto at the time did not appreciate this aspect and were essentially arguing for a ban on meta-analysis. Even about 2002/3, Iain Chalmers had to insist on the paper being referenced in Cochrane materials after it had been edited out a couple times by others.
Even today I try to emphasize compare (investigate differences) before any mention of combining and stress postulating and appraising commonness is primary. I originally posted to https://en.wikipedia.org/wiki/Meta-analysis to try to make that so around 2010 and largely succeeded (I think in the first sentence) to see it only slowly revised over the years to toady’s fourth sentence that starts “In addition to providing an estimate of the unknown common truth” that suggests its largely a secondary consideration.
> many people mistake “meta-analysis” and “systematic review” as synonyms
I did – but the vocabulary evolved in the late 1980s in UK and I was unaware of the distinctions – initially we had used meta-analysis as term closer to a synonym for today’s meta-research – see http://www.jameslindlibrary.org/labbe-ka-detsky-as-orourke-k-1987/
These challenges of vocabulary were actually written about by me and edited by Iain Chalmers here http://www.jameslindlibrary.org/articles/a-historical-perspective-on-meta-analysis-dealing-quantitatively-with-varying-study-results/ (copied below).
The continued debate referred to my deferring on this part to Iain “Meta-analysis may be, but is not necessarily, used as part of this process” when I wanted some like “Meta-analysis should be used if at all purposeful in this part of the process, but it may not always be purposeful.”
Believe or not, Systematic Review/Meta-analysis/Meta-Research is a relatively new but extremely important area of science – I hope you find it enjoyable.
Keith O’Rourke
By the early 1990s, terminology was becoming confusing, and Chalmers and Altman (1995) suggested that the term ‘meta-analysis’ should be restricted to the process of statistical synthesis considered in this commentary. This convention has now been adopted in some quarters. For example, the second edition of the BMJ publication Systematic Reviews is subtitled ‘meta-analysis in context’ (Egger et al. 2001), and the 4th edition of Last’s Dictionary of Epidemiology (Last 2001) gives definitions as follows:
Systematic Review: The application of strategies that limit bias in the assembly, critical appraisal, and synthesis of all relevant studies on a specific topic. Meta-analysis may be, but is not necessarily, used as part of this process.
Meta-Analysis: The statistical synthesis of the data from separate but similar, i.e. comparable studies, leading to a quantitative summary of the pooled results.
Just as debates seem likely to continue about the statistical methods used for meta-analysis, so also will debates continue about terminology. What is certain, however, is that we will continue to have to deal quantitatively with varying study results.
So nice to hear from you, Keith – and to hear about the context of that paper, the debates around terminology, and the history with the Wikipedia article. (Will definitely tackle that Wikipedia article some time – thanks for pointing that out.)
You might be interested to see that the process is still rolling out to other fields. Just today I saw this for the first time: an analysis of problems in tissue engineering research based on a systematic review.
Thanks again – and I’m glad you enjoyed the post.
This is a reasonably well researched and thought out paper. Unfortunately, some psychologists are taking it as evidence that there’s nothing to the published meta-analysis literature in the field or that meta-analyses should be avoided. I know you end with a conclusion that these are issues that should be addressed. However, the title, as it stands, doesn’t reflect that and naysayers may be less concerned about the last few lines of the paper. Perhaps just an “s” on “problem” would help.