Mixed Methods Appraisal Tool (Mmat)
Mixed Methods Appraisal Tool (Mmat)
Mixed Methods Appraisal Tool (Mmat)
VERSION 2018
User guide
Prepared by
Quan Nha HONGa, Pierre PLUYEa,, Sergi FÀBREGUESb, Gillian BARTLETTa, Felicity BOARDMANc,
Margaret CARGOd, Pierre DAGENAISe, Marie‐Pierre GAGNONf, Frances GRIFFITHSc, Belinda NICOLAUa,
Alicia O’CATHAINg, Marie‐Claude ROUSSEAUh, & Isabelle VEDELa
a
McGill University, Montréal, Canada; bUniversitat Oberta de Catalunya, Barcelona, Spain; cUniversity of Warwick, Coventry, England;
d
University of Canberra, Canberra, Australia; eUniversité de Sherbrooke, Sherbrooke, Canada; fUniversité Laval, Québec, Canada;
g
University of Sheffield, Sheffield, England; hInstitut Armand‐Frappier Research Centre, Laval, Canada
2
Part II: Explanations
Ethnography This criterion was considered important to add in the MMAT since there is only one category of criteria for qualitative studies
The aim of the study is to describe and interpret the shared cultural (compared to three for quantitative studies).
behaviour of a group of individuals. 1.2. Are the qualitative data collection methods adequate to address the research question?
Phenomenology Explanations
The study focuses on the subjective experiences and interpretations of a This criterion is related to data collection method, including data sources (e.g., archives, documents), used to address the
phenomenon encountered by individuals. research question. To judge this criterion, consider whether the method of data collection (e.g., in depth interviews and/or
group interviews, and/or observations) and the form of the data (e.g., tape recording, video material, diary, photo, and/or field
Narrative research notes) are adequate. Also, clear justifications are needed when data collection methods are modified during the study.
The study analyzes life experiences of an individual or a group. 1.3. Are the findings adequately derived from the data?
3
2. Quantitative Methodological quality criteria
randomized
controlled trials
Randomized controlled 2.1. Is randomization appropriately performed?
clinical trial: A clinical
study in which individual Explanations
participants are allocated In a randomized controlled trial, the allocation of a participant (or a data collection unit, e.g., a school) into the intervention or control group is based solely on chance.
to intervention or control Researchers should describe how the randomization schedule was generated. A simple statement such as ‘we randomly allocated’ or ‘using a randomized design’ is insufficient
groups by randomization to judge if randomization was appropriately performed. Also, assignment that is predictable such as using odd and even record numbers or dates is not appropriate. At minimum,
(intervention assigned by a simple allocation (or unrestricted allocation) should be performed by following a predetermined plan/sequence. It is usually achieved by referring to a published list of random
researchers). numbers, or to a list of random assignments generated by a computer. Also, restricted allocation can be performed such as blocked randomization (to ensure particular allocation
ratios to the intervention groups), stratified randomization (randomization performed separately within strata), or minimization (to make small groups closely similar with
Key references: Higgins respect to several characteristics). Another important characteristic to judge if randomization was appropriately performed is allocation concealment that protects assignment
and Green (2008); sequence until allocation. Researchers and participants should be unaware of the assignment sequence up to the point of allocation. Several strategies can be used to ensure
Higgins et al. (2016); allocation concealment such relying on a central randomization by a third party, or the use of sequentially numbered, opaque, sealed envelopes (Higgins et al., 2016).
Oxford Centre for 2.2. Are the groups comparable at baseline?
Evidence-based
Medicine (2016); Porta Explanations
et al. (2014) Baseline imbalance between groups suggests that there are problems with the randomization. Indicators from baseline imbalance include: “(1) unusually large differences
between intervention group sizes; (2) a substantial excess in statistically significant differences in baseline characteristics than would be expected by chance alone; (3) imbalance
in key prognostic factors (or baseline measures of outcome variables) that are unlikely to be due to chance; (4) excessive similarity in baseline characteristics that is not
compatible with chance; (5) surprising absence of one or more key characteristics that would be expected to be reported” (Higgins et al., 2016, p. 10).
2.3. Are there complete outcome data?
Explanations
Almost all the participants contributed to almost all measures. There is no absolute and standard cut-off value for acceptable complete outcome data. Agree among your team
what is considered complete outcome data in your field and apply this uniformly across all the included studies. For instance, in the literature, acceptable complete data value
ranged from 80% (Thomas et al., 2004; Zaza et al., 2000) to 95% (Higgins et al., 2016). Similarly, different acceptable withdrawal/dropouts rates have been suggested: 5% (de
Vet et al., 1997; MacLehose et al., 2000), 20% (Sindhu et al., 1997; Van Tulder et al., 2003) and 30% for a follow-up of more than one year (Viswanathan and Berkman, 2012).
2.4. Are outcome assessors blinded to the intervention provided?
Explanations
Outcome assessors should be unaware of who is receiving which interventions. The assessors can be the participants if using participant reported outcome (e.g., pain), the
intervention provider (e.g., clinical exam), or other persons not involved in the intervention (Higgins et al., 2016).
2.5 Did the participants adhere to the assigned intervention?
Explanations
To judge this criterion, consider the proportion of participants who continued with their assigned intervention throughout follow-up. “Lack of adherence includes imperfect
compliance, cessation of intervention, crossovers to the comparator intervention and switches to another active intervention.” (Higgins et al., 2016, p. 25).
4
3. Quantitative non-randomized studies Methodological quality criteria
Non-randomized studies are defined as any quantitative 3.1. Are the participants representative of the target population?
studies estimating the effectiveness of an intervention or
studying other exposures that do not use randomization to Explanations
allocate units to comparison groups (Higgins and Green, Indicators of representativeness include: clear description of the target population and of the sample (inclusion and exclusion criteria), reasons
2008). why certain eligible individuals chose not to participate, and any attempts to achieve a sample of participants that represents the target
population.
Common designs include (this list if not exhaustive): 3.2. Are measurements appropriate regarding both the outcome and intervention (or exposure)?
5
4. Quantitative descriptive studies Methodological quality criteria
Quantitative descriptive studies are “concerned with and 4.1. Is the sampling strategy relevant to address the research question?
designed only to describe the existing distribution of
variables without much regard to causal relationships or Explanations
other hypotheses” (Porta et al., 2014, p. 72). They are used Sampling strategy refers to the way the sample was selected. There are two main categories of sampling strategies: probability sampling
to monitoring the population, planning, and generating (involve random selection) and non-probability sampling. Depending on the research question, probability sampling might be preferable. Non-
hypothesis (Grimes and Schulz, 2002). probability sampling does not provide equal chance of being selected. To judge this criterion, consider whether the source of sample is
relevant to the target population; a clear justification of the sample frame used is provided; or the sampling procedure is adequate.
Common designs include the following single-group 4.2. Is the sample representative of the target population?
studies (this list if not exhaustive):
Explanations
Incidence or prevalence study without comparison There should be a match between respondents and the target population. Indicators of representativeness include: clear description of the target
group population and of the sample (such as respective sizes and inclusion and exclusion criteria), reasons why certain eligible individuals chose not
In a defined population at one particular time, what is to participate, and any attempts to achieve a sample of participants that represents the target population.
happening in a population, e.g., frequencies of factors 4.3. Are the measurements appropriate?
(importance of problems), is described (portrayed).
Explanations
Survey Indicators of appropriate measurements include: the variables are clearly defined and accurately measured, the measurements are justified and
“Research method by which information is gathered by appropriate for answering the research question; the measurements reflect what they are supposed to measure; validated and reliability tested
asking people questions on a specific topic and the data measures of the outcome of interest are used, variables are measured using ‘gold standard’, or questionnaires are pre-tested prior to data
collection procedure is standardized and well defined.” collection.
(Bennett et al., 2011, p. 3). 4.4. Is the risk of nonresponse bias low?
Explanations
The statistical analyses used should be clearly stated and justified in order to judge if they are appropriate for the design and research question,
and if any problems with data analysis limited the interpretation of the results.
6
5. Mixed methods studies Methodological quality criteria
Mixed methods (MM) research involves combining qualitative 5.1. Is there an adequate rationale for using a mixed methods design to address the research question?
(QUAL) and quantitative (QUAN) methods. In this tool, to be
considered MM, studies have to meet the following criteria (Creswell Explanations
and Plano Clark, 2017): (a) at least one QUAL method and one QUAN The reasons for conducting a mixed methods study should be clearly explained. Several reasons can be invoked such as to
method are combined; (b) each method is used rigorously in accordance enhance or build upon qualitative findings with quantitative results and vice versa; to provide a comprehensive and complete
to the generally accepted criteria in the area (or tradition) of research understanding of a phenomenon or to develop and test instruments (Bryman, 2006).
invoked; and (c) the combination of the methods is carried out at the 5.2. Are the different components of the study effectively integrated to answer the research question?
minimum through a MM design (defined a priori, or emerging) and the
integration of the QUAL and QUAN phases, results, and data. Explanations
Integration is a core component of mixed methods research and is defined as the “explicit interrelating of the quantitative and
Common designs include (this list if not exhaustive): qualitative component in a mixed methods study” (Plano Clark and Ivankova, 2015, p. 40). Look for information on how
qualitative and quantitative phases, results, and data were integrated (Pluye et al., 2018). For instance, how data gathered by both
Convergent design research methods was brought together to form a complete picture (e.g., joint displays) and when integration occurred (e.g.,
The QUAL and QUAN components are usually (but not necessarily) during the data collection-analysis or/and during the interpretation of qualitative and quantitative results).
concomitant. The purpose is to examine the same phenomenon by 5.3. Are the outputs of the integration of qualitative and quantitative components adequately interpreted?
interpreting QUAL and QUAN results (bringing data analysis together
at the interpretation stage), or by integrating QUAL and QUAN Explanations
datasets (e.g., data on same cases), or by transforming data (e.g., This criterion is related to meta-inference, which is defined as the overall interpretations derived from integrating qualitative and
quantization of qualitative data). quantitative findings (Teddlie and Tashakkori, 2009). Meta-inference occurs during the interpretation of the findings from the
integration of the qualitative and quantitative components, and shows the added value of conducting a mixed methods study
Sequential explanatory design rather than having two separate studies.
Results of the phase 1 - QUAN component inform the phase 2 - QUAL 5.4. Are divergences and inconsistencies between quantitative and qualitative results adequately addressed?
component. The purpose is to explain QUAN results using QUAL
findings. E.g., the QUAN results guide the selection of QUAL data Explanations
sources and data collection, and the QUAL findings contribute to the When integrating the findings from the qualitative and quantitative components, divergences and inconsistencies (also called
interpretation of QUAN results. conflicts, contradictions, discordances, discrepancies, and dissonances) can be found. It is not sufficient to only report the
divergences; they need to be explained. Different strategies to address the divergences have been suggested such as reconciliation,
Sequential exploratory design initiation, bracketing and exclusion (Pluye et al., 2009b). Rate this criterion ‘Yes’ if there is no divergence.
Results of the phase 1 - QUAL component inform the phase 2 - QUAN 5.5. Do the different components of the study adhere to the quality criteria of each tradition of the methods involved?
component. The purpose is to explore, develop and test an instrument
(or taxonomy), or a conceptual framework (or theoretical model). E.g., Explanations
the QUAL findings inform the QUAN data collection, and the QUAN The quality of the qualitative and quantitative components should be individually appraised to ensure that no important threats to
results allow a statistical generalization of the QUAL findings. trustworthiness are present. To appraise 5.5, use criteria for the qualitative component (1.1 to 1.5), and the appropriate criteria for
the quantitative component (2.1 to 2.5, or 3.1 to 3.5, or 4.1 to 4.5). The quality of both components should be high for the mixed
Key references: Creswell et al. (2011); Creswell and Plano Clark, methods study to be considered of good quality. The premise is that the overall quality of a mixed methods study cannot exceed
(2017); O'Cathain (2010) the quality of its weakest component. For example, if the quantitative component is rated high quality and the qualitative
component is rated low quality, the overall rating for this criterion will be of low quality.
7
Algorithm for selecting the study categories to rate in the MMAT*
*Adapted from National Institute for Health Care Excellence. (2012). Methods for the development of nice public health guidance. London: National Institute for Health and Care Excellence; and Scottish Intercollegiate
Guidelines Network. (2017). Algorithm for classifying study design for questions of effectiveness. Retrieved December 1, 2017, from http://www.sign.ac.uk/assets/study_design.pdf. 8
References
Abbott, A. (1998). The causal devolution. Sociological Methods & Research, 27(2), 148-181.
Bennett, C., Khangura, S., Brehaut, J. C., Graham, I. D., Moher, D., Potter, B. K., et al. (2011). Reporting guidelines for survey research: An analysis of published guidance and reporting practices. PLoS
Medicine, 8(8), e1001069.
Bryman, A. (2006). Integrating quantitative and qualitative research: How is it done? Qualitative Research, 6(1), 97-113.
Creswell, J. W. (2013a). Qualitative inquiry and research design: Choosing among five approaches (3rd ed.). Thousand Oaks, CA: SAGE Publications.
Creswell, J. W. (2013b). Research design: Qualitative, quantitative, and mixed methods approaches. Thousand Oaks, CA: SAGE Publications.
Creswell, J. W., Klassen, A. C., Plano Clark, V. L., Smith, K. C. (2011). Best practices for mixed methods research in the health sciences. Bethesda, MD: Office of Behavioral and Social Sciences
Research, National Institutes of Health. http://obssr.od.nih.gov/mixed_methods_research.
Creswell, J. W., & Plano Clark, V. (2017). Designing and conducting mixed methods research (3rd ed.). Thousand Oaks, CA: SAGE Publications.
Critical Appraisal Skills Programme. (2017). CASP checklists. Retrieved December 1, 2017, from http://www.casp-uk.net/casp-tools-checklists.
de Vet, H. C., de Bie, R. A., van der Heijden, G. J., Verhagen, A. P., Sijpkes, P., & Knipschild, P. G. (1997). Systematic reviews on the basis of methodological criteria. Physiotherapy, 83(6), 284-289.
Draugalis, J. R., Coons, S. J., & Plaza, C. M. (2008). Best practices for survey research reports: A synopsis for authors and reviewers. American Journal of Pharmaceutical Education, 72(1), Article 11.
Federal Committee on Statistical Methodology. (2001). Measuring and reporting sources of error in surveys. Washington DC: Statistical Policy Office, Office of Information and Regulatory Affairs,
Office of Management and Budget.
Grimes, D. A., & Schulz, K. F. (2002). Descriptive studies: What they can and cannot do. The Lancet, 359(9301), 145-149.
Higgins, J. P., & Green, S. (2008). Cochrane handbook for systematic reviews of interventions. Chichester, UK: Wiley Online Library.
Higgins, J. P. T., Sterne, J. A. C., Savović, J., Page, M. J., Hróbjartsson, A., Boutron, I., et al. (2016). A revised tool for assessing risk of bias in randomized trials. In Chandler, J., McKenzie, J., Boutron,
I. & Welch, V. (Eds.), Cochrane Methods. Cochrane Database of Systematic Reviews, Issue 10 (Suppl 1).
Hong, Q. N. (2018). Revision of the Mixed Methods Appraisal Tool (MMAT): A mixed methods study (Doctoral dissertation). Department of Family Medicine, McGill University, Montréal.
MacLehose, R. R., Reeves, B. C., Harvey, I. M., Sheldon, T. A., Russell, I. T., & Black, A. M. (2000). A systematic review of comparisons of effect sizes derived from randomised and non-randomised
studies. Health Technology Assessment, 4(34), 1-154.
Morgan, R., Sterne, J., Higgins, J., Thayer, K., Schunemann, H., Rooney, A., et al. (2017). A new instrument to assess Risk of Bias in Non-randomised Studies of Exposures (ROBINS-E): Application to
studies of environmental exposure. Abstracts of the Global Evidence Summit, Cape Town, South Africa. Cochrane Database of Systematic Reviews 2017, Issue 9 (Suppl 1).
https://doi.org/10.1002/14651858.CD201702.
O'Cathain, A. (2010). Assessing the quality of mixed methods research: Towards a comprehensive framework. In Tashakkori, A. & Teddlie, C. (Eds.), Handbook of Mixed methods in social and
behavioral research (pp. 531-555). Thousand Oaks, CA: SAGE Publications.
Oxford Centre for Evidence-based Medicine. (2016). Levels of evidence. Retrieved February 19, 2018, from https://www.cebm.net/2016/05/ocebm-levels-of-evidence/.
Pace, R., Pluye, P., Bartlett, G., Macaulay, A. C., Salsberg, J., Jagosh, J., et al. (2012). Testing the reliability and efficiency of the pilot Mixed Methods Appraisal Tool (MMAT) for systematic mixed
studies review. International Journal of Nursing Studies, 49(1), 47-53.
Plano Clark, V. L., & Ivankova, N. V. (2015). Mixed methods research: A guide to the field. Thousand Oaks, CA: SAGE Publications.
Pluye, P., Gagnon, M. P., Griffiths, F., Johnson-Lafleur, J. (2009a). A scoring system for appraising mixed methods research, and concomitantly appraising qualitative, quantitative and mixed methods
primary studies in mixed studies reviews. International Journal of Nursing Studies, 46(4), 529-546.
Pluye, P., Grad, R. M., Levine, A., & Nicolau, B. (2009b). Understanding divergence of quantitative and qualitative data (or results) in mixed methods studies. International Journal of Multiple Research
Approaches, 3(1), 58-72.
Pluye, P., Garcia Bengoechea, E., Granikov, V., Kaur, N., & Tang, D. L. (2018). A world of possibilities in mixed methods: Review of the combinations of strategies used to integrate the phases, results,
and qualitative and quantitative data. International Journal of Multiple Research Approaches, 10(1), 41-56.
Porta, M. S., Greenland, S., Hernán, M., dos Santos Silva, I., Last, J. M. (2014). A dictionary of epidemiology. New York: Oxford University Press.
Sandelowski, M. (2010). What's in a name? Qualitative description revisited. Research in Nursing and Health, 33(1), 77-84.
9
Schwandt, T. A. (2015). The SAGE dictionary of qualitative inquiry. Thousand Oaks, CA: SAGE Publications.
Sindhu, F., Carpenter, L., & Seers, K. (1997). Development of a tool to rate the quality assessment of randomized controlled trials using a Delphi technique. Journal of Advanced Nursing, 25(6), 1262-
1268.
Sterne, J. A., Hernán, M. A., Reeves, B. C., Savović, J., Berkman, N. D., Viswanathan, M., et al. (2016). ROBINS-I: A tool for assessing risk of bias in non-randomised studies of interventions. British
Medical Journal, 355(i4919).
Teddlie, C., & Tashakkori, A. (2009). Foundations of mixed methods research: Integrating quantitative and qualitative approaches in the social and behavioral sciences. Thousand Oaks, CA: SAGE
Publications.
Thomas, B. H., Ciliska, D., Dobbins, M., & Micucci, S. (2004). A process for systematically reviewing the literature: Providing the research evidence for public health nursing interventions. Worldviews
on Evidence-Based Nursing, 1(3), 176-184.
Van Tulder, M., Furlan, A., Bombardier, C., Bouter, L., & Editorial Board of the Cochrane Collaboration Back Review Group. (2003). Updated method guidelines for systematic reviews in the Cochrane
Collaboration Back Review Group. Spine, 28(12), 1290-1299.
Viswanathan, M., & Berkman, N. D. (2012). Development of the RTI item bank on risk of bias and precision of observational studies. Journal of Clinical Epidemiology, 65(2), 163-178.
Wells, G., Shea, B., O’connell, D., Peterson, J., Welch, V., Losos, M., et al. (2000). The Newcastle-Ottawa Scale (NOS) for assessing the quality of nonrandomised studies in meta-analyses. Retrieved
April 16, 2016, from http://www.ohri.ca/programs/clinical_epidemiology/nosgen.pdf.
Zaza, S., Wright-De Agüero, L. K., Briss, P. A., Truman, B. I., & Hopkins, D. P. (2000). Data collection instrument and procedure for systematic reviews in the guide to community preventive services.
American Journal of Preventive Medicine, 188(Suppl 1), 44-74.
10