1410 Secondary
1410 Secondary
1410 Secondary
Secondary research refers to research conducted using data from other studies. When conducting secondary research, individual patients are not
recruited but eligible studies are ‘recruited’. Two important types of secondary research are 1] Narrative reviews 2] Systematic reviews
Narrative reviews:
These reviews are usually carried out by experts in the field of study and are more or less aligned with the expert’s own opinion on the issue supported
by selected research evidence. These reviews are generally broad without a narrow or specific clinical question and are subjective. They carry a
higher risk of being prone to bias than systematic reviews. Literature is not searched methodically but rather in a random manner to suit one’s flow of
argument. Because of the above, narrative reviews cannot be replicated (unless repeated by the same author!). Nevertheless narrative reviews can
provide a good introduction to a topic, stimulate interest and controversies and can even collate existing data for generating new hypotheses.
Systematic reviews:
Systematic reviews follow rigorous steps in identifying relevant
literature, appraising individual studies and analysing suitable
data to synthesise a conclusion. The characteristics of a
systematic review are
1. Study a focused narrow question
2. Comprehensive & specific data collection (see table)
3. Uniform criteria for study selection (see table)
4. Quantitative synthesis of data (may or may not be
possible)
The 4th feature is optional as not all data are ‘combinable’ to
produce a final common result. Also in some situations good
quality data may not be available to combine and synthesise a
concluding result. If the statistical combination is possible, then
such procedure of quantitative synthesis of individual study
data is called a meta-analysis.
Individual trials’ quality: If the primary studies included have poor quality, then even if sound methodological and statistical procedures are followed
the outcome may be meaningless. This is called GIGO principle of ‘garbage in, garbage out’!!
The quality of individual trials can be assessed by considering
1. The nature of the patient sample,
2. Outcomes studied,
3. Length of follow up and
4. Comparability (e.g. dosage) of treatment and
5. Methodological factors such as
1. Adequacy of sample randomisation, 2.adequate concealment and control of intervention, 3. Analysis with the intention to treat principles, and 4. Having an
objective, blinded, outcome assessment
© SPMM Course
A PROPER LITERATURE SEARCH MUST HAVE FOLLOWING The inclusion criteria for studies in a systematic
CHARACTERISTICS: review should consider
1. Using multiple databases – no just limiting to MEDLINE only or CINAHL (1) Types of study designs to include
only.
(2) Types of subjects to include
2. Cross checking the references list of each individual study retrieved by
direct search (3) Types of publications
3. Hand searching for materials unidentified online. (4) Language restrictions, if any
4. Approaching experts to comment on any missing studies.
(5) Types of interventions
5. Identifying grey literature – i.e. the unpublished literature such as
conference abstracts, presentations and posters. (6) Time frame for included studies
Meta-analysis:
• A meta-analysis is most often used to assess the clinical effectiveness of healthcare interventions.
• Hence RCTs are the most frequently combined studies for a meta-analysis though it is possible to combine other study designs.
• A meta-analysis may either use summary data or individual patient data. For
Quality
of
a
meta-‐analysis
or
systematic
review
depends
on
the
example, if one wants to do a meta-analysis of 3 studies each having 50 patients
enrolled, using summary method your N will be 3. In case if one is following following
factors
individual patient data method this requires contacting the authors of each of the •A comprehensive literature search ( see below)
3 trials and requesting the data of the subjects. Though this is ideal, the inability •Clearly defined eligibility criteria for the studies to be included ( see
of investigators to supply data, and the increased costs associated with the below)
retrieval of individual patient data make researchers carry out summary data •Clearly defined & strictly adhered protocol.
method more often. •Predetermined criteria related to the quality of trials
• The four basic steps for conducting a meta-analysis include •Assessing the quality of a study can be difficult since the information
(1) Literature search, reported is often inadequate for this purpose. In these cases, a thorough
(2) Establishing criteria for including and excluding studies sensitivity analysis must be conducted.
(3) Recording of data from the individual studies, and •Discussion and analysis of statistical and clinical heterogeneity.
(4) Statistical analysis of the data.
Heterogeneity can occur for clinical reasons in trials. For example, even if the same outcome / end result is considered (e.g. increase in insight), the
interventions may be diverse, such as providing psychoeducation sessions, leaflets, videotapes, audiocassettes or talking to the family. Though the
results can be combined theoretically (all odds risk differences having weighted mean) this would not be meaningful. The interventions are so
different that combining them does not make clinical sense. This is an example of clinical heterogeneity. Other circumstances that may give rise to
clinical heterogeneity include differences in selection of patients, the severity of disease and dose or duration of treatment. Clinical heterogeneity is
describable but not measurable. (Retrieved from http://www.bmj.com/content/334/7584/94)
Methodological Heterogeneity:
This refers to heterogeneity resulting from the differential use of study methodology. These may lead to different conclusions in different studies, even
though the clinical characteristics are the same. Methodological heterogeneity is describable but does not need quantification.
Statistical heterogeneity
In a systematic review, individual studies may purport to measure the same outcome but often report results that are not consistent with each other.
Some trials may show a benefit while others show a lack of response, or the trials may be inconsistent in the size of the observed difference between the
placebo and active treatment arms. This variation is called statistical heterogeneity. A homogenous sample refers to a set of individual studies that have
comparable outcomes without much variation (i.e they do not disagree greatly among each other). Heterogeneity refers to the presence of significant
variation among the individual studies in a sample.
© SPMM Course
§ Heterogeneity may be judged graphically or be measured statistically.
§ Forest Plot & L'Abbé plot can be used to explore the inconsistency of studies visually. Forest plot is described elsewhere (see blobbogram).
§ L’Abbe plot is nothing but a modified scatter plot wherein CER is plotted aganst EER from individual trials included in the meta-analysis.
Trials that report experimental treatment to be superior to the control arm (EER > CER) will be in the upper left of the plot, between the y-axis
and the diagonal line of equality. If experimental is no better than control then the point will fall on the line of equality (EER = CER), and if
control is better than experimental then the point will be in the lower right of the plot, between the x-axis and the line of equality (EER < CER).
This provides a visually clear idea of how heterogeneous the results are.
àL’Abbe plot:
Publication bias:
• This refers to the tendency for authors to submit and/or editors to publish only those studies that yield statistically significant results (i.e. ‘positive’
and rejecting the null hypothesis).
• There are several methods available to diagnose the presence of publication bias in the literature. Most of these methods operate on the basis that
small studies are more susceptible to publication bias and may, consequently, show larger treatment effects.
• Application of tests for publication bias in the presence of heterogeneity (described below) is not valid, and may lead to false-positive claims for
publication bias (Ioannidis, 2007).
• When all available studies are equally large (i.e., have similar precision), the tests are not meaningful.
Funnel plot: The most common method to detect publication bias is to investigate the presence of asymmetry in (inverted) funnel plots. A funnel plot
shows the relation between the effect size and precision of the individual studies in a meta-anaysis. Small studies are more likely to remain unpublished
if their results are nonsignificant or unfavourable, and larger studies get published regardless prducing an asymmetrical funnel-plot. Although visual
inspection of funnel plots is often unreliable, several statistical tests can be used to quantify
the asymmetry.
Failsafe N: Meta-analytic results could be wrong, due to non-significant studies that remained
in researchers’ ‘file drawers’ and not published. One can speculate whether the inclusion of
these studies would nullify the statistical significance of an observed mean effect in a meta-
analysis. Rosenthal & Cooper developed a formula to calculate the number of zero-effect
studies (no effect) that would be required to nullify the mean effect seen in a meta-analysis.
This number was termed the failsafe N or file-drawer number.
Other tests for publication bias:
Either linear regression method or correlation method (Tau)
can be used to quantify the publication bias. Details of such methods are beyond the scope of
MRCPsych exams.
Cumulative meta-analysis: This can be used to assess the potential impact of publication
bias in tilting or nullifying an effect. A cumulative meta-analysis is done in the following
manner:
o The studies are sorted from largest to smallest (or from most to least precise).
o A meta-analysis is initially run with one study, then repeated with a second study
added, then a third, and so on;
o A forest plot/blobbogram (see elsewhere in the notes) is plotted as one goes along the process of accumulation. Here the first row will show
the effect based on one study, the second row shows the cumulative effect based on two àFunnel plot:
studies and so on.
o If the point estimate stabilizes based on the initially added larger studies, then there is no Note that the funnel is well formed in this example. X axis of
evidence that the smaller studies are producing a biased overall effect. funnel plot carries a measure of effect size (OR, log OR,
o If, however, the point estimate shifts when the smaller studies are added, this can be Cohen’s d etc). Y axis carries a measure of precision such as
considered as an evidence of bias and one can also see the direction of the biased effect. sample size, inverse of standard error (1/SE) etc. (From
Cochrane database training section)
© SPMM Course
The trim-and-fill procedure: This was developed by Duval and Tweedie (2000). This helps to assess whether the effect would change if the bias were to
be removed. Trim and fill procedures is a kind of sensitivity analysis where missing studies are imputed and added to the analysis, and then effect size
is recomputed.
(Other biases that can occur include location bias – studies not being located due to citation habits, the database used, key terms used or multiple
replications of same data.
Inclusion bias refers to the reviewers’ tendency to include studies they agree with.)
Sensitivity analysis: This refers to the analysis of the extent to which an outcome derived from research will change when ‘hypothetical data’ are
introduced. One always wonders if an established result could be different if certain aspects of a research process were different. Such an exploration
of ‘what ifs’ is called sensitivity analysis (Greenhalgh, 2001).This applies to mainly to economic analysis and meta-analysis. In a meta-analysis, there are
various nodal points where such hypothetical data can be introduced. For example
1. What if there was significant heterogeneity (or absence of it) – how will this affect the outcome? One can do analysis in both random and
fixed effects models.
2. Factors determining methodological quality e.g. numbers allocated in either arm, type of outcome measure used, case definition used, etc.
can be manipulated for sensitivity analyses.
3. Publication bias can be examined using sensitivity analysis methods e.g. failsafe N.
Blobbograms:
o A blobbogram or forest plot presents the effect (point estimate) from each individual study as a blob or square (the measured effect), with a
horizontal line (usually the 95% confidence interval, indicating the precision) across the blob.
o The size of the blob or square Forest plot: Things to note in a forest plot
reflects the amount of information in
that individual study (often the 1. The combined /pooled effect size (given by various summary statistics including OR, RR, standardised mean
sample size); differences, Cohen’s d etc)
o The length of the horizontal line 2. Confidence interval of individual point (line width) estimates and combined estimate (lozenge width)
represents the degee of uncertainty 3. Weighting assigned to studies (rectangle size)
of the estimated treatment effect for 4. Heterogeneity of results (absence of heterogeneity indicated by vertical linearity of rectangles or non significant chi
that study. square test for heterogeneity)
American Journal of Psychiatry 2003; 160:1919–1928
o A mid-vertical line is often
constructed to represent null effect for the differences in outcome; on either side lies the area favouring intervention A or B. When the blob or
square lies on one side of this vertical line, this means that the outcome of that specific favoured the intervention marked on that side. If an
individual study result is statisticsally significant then the horizontal lines representing uncertaininty will lie entirely on the same side of the
blob, not touching or cross the midvertical line.
o A final synthesis of individual studies is represented in the form of a diamond lozenge. The position of the lozenge depends upon whether the
intervention rates favourably or not at the end of the meta-analysis.
o The following figure is a forest plot
© SPMM Course
Effect size:
§ Effect size measures for dichotomous variables: In meta-analyses, relative risk and odds ratio have the advantage in comparison with the absolute
risk that they depend less on the baseline risk. OR or RR are appropriate as an effect size measure when both variables are binary (cured vs. not
cured). Conventionally, the odds ratio is commonly used in reporting treatment effects. OR is more or less equal to relative risk and so an odds
ratio of 3 can be interpreted as ‘the outcome of interest happens
about thrice as often in the intervention group as in the control
group’. But this is not accurate (simply equating OR and RR) as this
will tend to overestimate the effect of treatment. But the level of this
overstatement is almost always quite small, especially when the
experimental and control event rates are less than 20%. However,
for frequent events (approximately over 20%), the odds ratio turns
out significantly greater than the relative risk, and falsely equating
relative risk and odds ratio, then leads to an overestimation of the
risk. For this reason, the relative risk is generally used in Cochrane
reviews.
§ Effect sizes for continuous variables: There are basically two
effect size measures for continuous variables: Simple difference
between the mean values (DM) keeps the original units and simply
is the mean value of group X minus the mean value of group Y.
Standardized difference between the mean values (SMD): This is
carried out by dividing the DM by the pooled standard deviation of
the two groups in the basic formula: SMD = (mean value of group X )
mean value of group Y) ! pooled standard deviation of the two
groups. Various formulas for the calculation of SMDs exist such as
e.g. Cohen’s d or Hedges' g, the latter avoiding a bias in Cohen’s d,
seen with small number of subjects.
Statistical methods for fixed vs. random effects:
Mantel-Haenszel ( see SPMM notes on advanced statistics) OR, Rate ratios, Risk ratios (RR)
Fixed effects
Peto Approximates OR
None of the both above mentioned fixed effects methods allow for confounding to be assessed – hence they are not useful if confounding is suspected due to methodological issues. They are
chi-square like tests, which need 2X2 tables for each individual study.
© SPMM Course
A meta-analysis may not always give results that equate with large mega-trials that may be conducted later. For example google ‘Lessons from an
effective, safe, simple intervention that was not’ (use of intravenous magnesium after heart attacks).
The QUOROM guidelines: Moher D et al. Improving the quality of reports of meta-analyses of
randomized controlled trials: the QUOROM statement. Lancet. 1999;354:1896-1900
Hunter JE, Schmidt FL. Fixed effects vs. random effects meta-analysis models: implications for cumulative
research knowledge. International Journal of Selection and Assessment 8:275-292, 2000
Ioaniidis, JPA & Trikalinos TA. The appropriateness of asymmetry tests for publication bias in meta-
analyses: a large survey. CMAJ 2007; 176 (8). doi:10.1503/cmaj.060410
DISCLAIMER: This material is developed from various revision notes assembled while preparing for MRCPsych exams. The content is periodically updated with excerpts from various
published sources including peer-reviewed journals, websites, patient information leaflets and books. These sources are cited and acknowledged wherever possible; due to the
structure of this material, acknowledgements have not been possible for every passage/fact that is common knowledge in psychiatry. We do not check the accuracy of drug-
related information using external sources; no part of these notes should be used as prescribing information.
© SPMM Course