In statistics, a meta-analysis is a sub-type of systematic reviews that combines the results of several studies that address a set of related research hypotheses. The first meta-analysis was performed by Karl Pearson in 1904, in an attempt to overcome the problem of reduced statistical power in studies with small sample sizes; analyzing the results from a group of studies can allow more accurate data analysis.
Although meta-analysis is widely used in epidemiology and evidence-based medicine today, a meta-analysis of a medical treatment was not published until 1955. In the 1970s, more sophisticated analytical techniques were introduced in educational research, starting with the work of Gene V. Glass, Frank L. Schmidt, and John E. Hunter.
The online Oxford English Dictionary lists the first usage of the term in the statistical sense as 1976 by Glass. The statistical theory surrounding meta-analysis was greatly advanced by the work of Nambury S. Raju, Larry V. Hedges, Ingram Olkin, John E. Hunter, and Frank L. Schmidt.
Because the results from different studies investigating different independent variables are measured on different scales, the dependent variable in a meta-analysis is some standardized measure of effect size. To describe the results of comparative experiments the usual effect size indicator is the standardized mean difference (d) which is the standard score equivalent to the difference between means, or an odds ratio if the outcome of the experiments is a dichotomous variable (success versus failure). A meta-analysis can be performed on studies that describe their findings in correlation coefficients, as for example, studies of the correlation between familial relationships and intelligence. In these cases, the correlation itself is the indicator of the effect size.
The method is not restricted to situations in which one or more variables is defined as "dependent." For example, a meta-analysis could be performed on a collection of studies each of which attempts to estimate the incidence of left-handedness in various groups of people.
Researchers should be aware that variations in sampling schemes can introduce heterogeneity to the result, which is the presence of more than one intercept in the solution. For instance, if some studies used 30mg of a drug, and others used 50mg, then we would plausibly expect two clusters to be present in the data, each varying around the mean of one dosage or the other. This can be modelled using a "random effects model."
Results from studies are combined using different approaches. One approach frequently used in meta-analysis in health care research is termed 'inverse variance method'. The average effect size across all studies is computed as a weighted mean, whereby the weights are equal to the inverse variance of each studies' effect estimator. Larger studies and studies with less random variation are given greater weight than smaller studies. Other common approaches include the Mantel Haenszel method and the Peto method.
Note, however that Mantel Haenszel analysis and Peto analysis introduce their own biases and distortions of the data results.
A recent approach to studying the influence that weighting schemes can have on results has been proposed through the construct of gravity, which is a special case of combinatorial meta analysis.
Modern meta-analysis does more than just combine the effect sizes of a set of studies. It can test if the studies' outcomes show more variation than the variation that is expected because of sampling different research participants. If that is the case, study characteristics such as measurement instrument used, population sampled, or aspects of the studies' design are coded. These characteristics are then used as predictor variables to analyze the excess variation in the effect sizes. Some methodological weaknesses in studies can be corrected statistically. For example, it is possible to correct effect sizes or correlations for the downward bias due to measurement error or restriction on score ranges.
Meta analysis leads to a shift of emphasis from single studies to multiple studies. It emphasises the practical importance of the effect size instead of the statistical significance of individual studies. This shift in thinking has been termed Metaanalytic thinking.
The results of a meta-analysis are often shown in a forest plot.
Assessing the quality of a trial by only using the published report may lead to inaccurate conclusions.[3]
A weakness of the method is that sources of bias are not controlled by the method. A good meta-analysis of badly designed studies will still result in bad statistics. Robert Slavin has argued that only methodologically sound studies should be included in a meta-analysis, a practice he calls 'best evidence meta-analysis'. Other meta-analysts would include weaker studies, and add a study-level predictor variable that reflects the methodological quality of the studies to examine the effect of study quality on the effect size. Another weakness of the method is the heavy reliance on published studies, which may increase the effect as it is very hard to publish studies that show no significant results. This publication bias or "file-drawer effect" (where non-significant studies end up in the desk drawer instead of in the public domain) should be seriously considered when interpreting the outcomes of a meta-analysis. Because of the risk of publication bias, many meta-analyses now include a "failsafe N" statistic that calculates the number of studies with null results that would need to be added to the meta-analysis in order for an effect to no longer be reliable.
The small study effect is the observation that small studies tend to report more positive results.[4][5][6] This is especially a threat when the original studies in a meta-analysis are less than 50 patients in size.[7]
Consistency can be statistically tested using either the Cochran's Q or I2.[8][9] The I2 is the "percentage of total variation across studies that is due to heterogeneity rather than chance."[8] These numbers are usually displayed for each group of studies on a Forest plot.
In interpreting of the Cochran's Q, heterogeneity exists if its p-value is < 0.05 or possibly if < 0.10[10][11].
The following has been proposed for interpreting I2:[8]
Individual patient data meta-analysis can be done with a one-stage or two-stage approach. The two-stage approach, which does not require true pooling of individual patient data, can yield very similar results if confounders or modulators are statistically controlled for[27].
Grounded thematic analysis can guide the collection of qualitative or narrative observations[35]. This approach can also be used in a positive deviance approach to quality improvement assessments[36]
Pragmatism. "what to include is not self-evident. The reviewer must be guided by what will be most useful to the intended audience(s), for example, what is likely to promote sense making.
Pluralism. "the topic should be illuminated from multiple angles and perspectives"
Historicity. "research traditions are often best described as they unfolded over time, highlighting significant individual scientists, events and discoveries which shaped the tradition."
Contestation. "'conflicting data' from different research traditions should be examined to generate higher-order insights"
Reflexivity. "throughout the review, reviewers must continually reflect, individually and as a team, on the emerging findings."
Peer review. "emerging findings should be presented to an external audience"
The conclusions of meta-analyses may be mitigated by research published after the search date of the meta-analysis. This may occur by the time the meta-analysis has been published.[40][41] Strategies have been developed for updating meta-analyses.[42]
↑Thijs V, Lemmens R, Fieuws S. Network meta-analysis: simultaneous meta-analysis of common antiplatelet regimens after transient ischaemic attack or stroke. ur Heart J. 2008 May;29(9):1086-92. Epub 2008 Mar 17. PMID 18349026
↑ 31.031.1Glaser, Barney G.; Strauss, Anselm L. (1967). The discovery of grounded theory : strategies for qualitative research. Chicago. ISBN978-0-202-30028-3. OCLC253912.
↑ 32.032.1Corbin, Juliet M.; Strauss, Anselm L. (2015). Basics of qualitative research : techniques and procedures for developing grounded theory (4 ed.). Thousand Oaks, California. ISBN1-4129-9746-1. OCLC898334340.
↑Patton, Michael Quinn (2015). Qualitative research & evaluation methods : integrating theory and practice. Thousand Oaks, California. ISBN1-4129-7212-4. OCLC890080219.
↑Wong G, Greenhalgh T, Westhorp G, Pawson R (2014). "Development of methodological guidance, publication standards and training materials for realist and meta-narrative reviews: the RAMESES (Realist And Meta-narrative Evidence Syntheses – Evolving Standards) project". Health Services and Delivery Research. doi:10.3310/hsdr02300. PMID25642521.CS1 maint: Multiple names: authors list (link)
↑Shojania KG, Sampson M, Ansari MT, Ji J, Doucette S, Moher D (2007). "How quickly do systematic reviews go out of date? A survival analysis". Ann. Intern. Med. 147 (4): 224–33. PMID17638714. Unknown parameter |month= ignored (help)CS1 maint: Multiple names: authors list (link)