Last update: August 1
st
, 2018

Preparedby
QuanNhaHONG
a
,PierrePLUYE
a,
,SergiFÀBREGUES
b
,GillianBARTLETT
a
,FelicityBOARDMAN
c
,
MargaretCARGO
d
,PierreDAGENAIS
e
,MariePierreGAGNON
f
,FrancesGRIFFITHS
c
,BelindaNICOLAU
a
,
AliciaO’CATHAIN
g
,MarieClaudeROUSSEAU
h
,&IsabelleVEDEL
a

a
McGillUniversity,Montréal,Canada;
b
UniversitatObertadeCatalunya,Barcelona,Spain;
c
UniversityofWarwick,Coventry,England;
d
UniversityofCanberra,Canberra,Australia;
e
UniversitédeSherbrooke,Sherbrooke, Canada;
f
UniversitéLaval,Québec,Canada;
g
UniversityofSheffield,Sheffield,England;
h
InstitutArmandFrappierResearchCentre,Laval,Canada
MIXEDMETHODSAPPRAISALTOOL(MMAT)
VERSION2018
Userguide
For dissemination, application, and feedback: Please contact mixed.methods.appraisal.too[email protected]
For more information: http://mixedmethodsappraisaltoolpublic.pbworks.com/ 1
What is the MMAT?
The MMAT is a critical appraisal tool that is designed for the appraisal stage of
systematic mixed studies reviews, i.e., reviews that include qualitative, quantitative and
mixed methods studies. It permits to appraise the methodological quality of five
categories to studies: qualitative research, randomized controlled trials, non-randomized
studies, quantitative descriptive studies, and mixed methods studies.
How was the MMAT developed?
The MMAT was developed in 2006 (Pluye et al., 2009a) and was revised in 2011 (Pace
et al., 2012). The present version 2018 was developed on the basis of findings from a
literature review of critical appraisal tools, interviews with MMAT users, and an e-
Delphi study with international experts (Hong, 2018). The MMAT developers are
continuously seeking for improvement and testing of this tool. Users’ feedback is always
appreciated.
What the MMAT can be used for?
The MMAT can be used to appraise the quality of empirical studies, i.e., primary
research based on experiment, observation or simulation (Abbott, 1998; Porta et al.,
2014). It cannot be used for non-empirical papers such as review and theoretical papers.
Also, the MMAT allows the appraisal of most common types of study methodologies
and designs. However, some specific designs such as economic and diagnostic accuracy
studies cannot be assessed with the MMAT. Other critical appraisal tools might be
relevant for these designs.
What are the requirements?
Because critical appraisal is about judgment making, it is advised to have at least two
reviewers independently involved in the appraisal process. Also, using the MMAT
requires experience or training in these domains. For instance, MMAT users may be
helped by a colleague with specific expertise when needed.
How to use the MMAT?
This document comprises two parts: checklist (Part I) and explanation of the criteria
(Part II).
1. Respond to the two screening questions. Responding ‘No’ or ‘Can’t tell’ to one or
both questions might indicate that the paper is not an empirical study, and thus
cannot be appraised using the MMAT. MMAT users might decide not to use these
questions, especially if the selection criteria of their review are limited to empirical
studies.
2. For each included study, choose the appropriate category of studies to appraise. Look
at the description of the methods used in the included studies. If needed, use the
algorithm at the end of this document.
3. Rate the criteria of the chosen category. For example, if the paper is a qualitative
study, only rate the five criteria in the qualitative category. The ‘Can’t tell’ response
category means that the paper do not report appropriate information to answer ‘Yes’
or ‘No’, or that report unclear information related to the criterion. Rating ‘Can’t tell’
could lead to look for companion papers, or contact authors to ask more information
or clarification when needed. In Part II of this document, indicators are added for
some criteria. The list is not exhaustive and not all indicators are necessary. You
should agree among your team which ones are important to consider for your field
and apply them uniformly across all included studies from the same category.
How to score?
It is discouraged to calculate an overall score from the ratings of each criterion. Instead,
it is advised to provide a more detailed presentation of the ratings of each criterion to
better inform the quality of the included studies. This may lead to perform a sensitivity
analysis (i.e., to consider the quality of studies by contrasting their results). Excluding
studies with low methodological quality is usually discouraged.
How to cite this document?
Hong QN, Pluye P, Fàbregues S, Bartlett G, Boardman F, Cargo M, Dagenais P, Gagnon
M-P, Griffiths F, Nicolau B, O’Cathain A, Rousseau M-C, Vedel I. Mixed Methods
Appraisal Tool (MMAT), version 2018. Registration of Copyright (#1148552), Canadian
Intellectual Property Office, Industry Canada.
2
Part I: Mixed Methods Appraisal Tool (MMAT), version 2018
Category of study
designs
Methodological quality criteria
Responses
Yes No Can’t tell Comments
Screening questions
(for all types)
S1. Are there clear research questions?
S2. Do the collected data allow to address the research questions?
Further appraisal may not be feasible or appropriate when the answer is ‘No’ or ‘Can’t tell’ to one or both screening questions.
1. Qualitative 1.1. Is the qualitative approach appropriate to answer the research question?
1.2. Are the qualitative data collection methods adequate to address the research question?
1.3. Are the findings adequately derived from the data?
1.4. Is the interpretation of results sufficiently substantiated by data?
1.5. Is there coherence between qualitative data sources, collection, analysis and interpretation?
2. Quantitative
randomized controlled
trials
2.1. Is randomization appropriately performed?
2.2. Are the groups comparable at baseline?
2.3. Are there complete outcome data?
2.4. Are outcome assessors blinded to the intervention provided?
2.5 Did the participants adhere to the assigned intervention?
3. Quantitative non-
randomized
3.1. Are the participants representative of the target population?
3.2. Are measurements appropriate regarding both the outcome and intervention (or exposure)?
3.3. Are there complete outcome data?
3.4. Are the confounders accounted for in the design and analysis?
3.5. During the study period, is the intervention administered (or exposure occurred) as intended?
4. Quantitative
descriptive
4.1. Is the sampling strategy relevant to address the research question?
4.2. Is the sample representative of the target population?
4.3. Are the measurements appropriate?
4.4. Is the risk of nonresponse bias low?
4.5. Is the statistical analysis appropriate to answer the research question?
5. Mixed methods 5.1. Is there an adequate rationale for using a mixed methods design to address the research question?
5.2. Are the different components of the study effectively integrated to answer the research question?
5.3. Are the outputs of the integration of qualitative and quantitative components adequately interpreted?
5.4. Are divergences and inconsistencies between quantitative and qualitative results adequately addressed?
5.5. Do the different components of the study adhere to the quality criteria of each tradition of the methods involved?
3
Part II: Explanations
1. Qualitative studies Methodological quality criteria
“Qualitative research is an approach for exploring and understanding the
meaning individuals or groups ascribe to a social or human problem”
(Creswell, 2013b, p. 3).
Common qualitative research approaches include (this list if not
exhaustive):
Ethnography
The aim of the study is to describe and interpret the shared cultural
behaviour of a group of individuals.
Phenomenology
The study focuses on the subjective experiences and interpretations of a
phenomenon encountered by individuals.
Narrative research
The study analyzes life experiences of an individual or a group.
Grounded theory
Generation of theory from data in the process of conducting research (data
collection occurs first).
Case study
In-depth exploration and/or explanation of issues intrinsic to a particular
case. A case can be anything from a decision-making process, to a person,
an organization, or a country.
Qualitative description
There is no specific methodology, but a qualitative data collection and
analysis, e.g., in-depth interviews or focus groups, and hybrid thematic
analysis (inductive and deductive).
Key references: Creswell (2013a); Sandelowski (2010); Schwandt (2015)
1.1. Is the qualitative approach appropriate to answer the research question?
Explanations
The qualitative approach used in a study (see non-exhaustive list on the left side of this table) should be appropriate for the
research question and problem. For example, the use of a grounded theory approach should address the development of a
theory and ethnography should study human cultures and societies.
This criterion was considered important to add in the MMAT since there is only one category of criteria for qualitative studies
(compared to three for quantitative studies).
1.2. Are the qualitative data collection methods adequate to address the research question?
Explanations
This criterion is related to data collection method, including data sources (e.g., archives, documents), used to address the
research question. To judge this criterion, consider whether the method of data collection (e.g., in depth interviews and/or
group interviews, and/or observations) and the form of the data (e.g., tape recording, video material, diary, photo, and/or field
notes) are adequate. Also, clear justifications are needed when data collection methods are modified during the study.
1.3. Are the findings adequately derived from the data?
Explanations
This criterion is related to the data analysis used. Several data analysis methods have been developed and their use depends on
the research question and qualitative approach. For example, open, axial and selective coding is often associated with grounded
theory, and within- and cross-case analysis is often seen in case study.
1.4. Is the interpretation of results sufficiently substantiated by data?
Explanations
The interpretation of results should be supported by the data collected. For example, the quotes provided to justify the themes
should be adequate.
1.5. Is there coherence between qualitative data sources, collection, analysis and interpretation?
Explanations
There should be clear links between data sources, collection, analysis and interpretation.
4
2. Quantitative
randomized
controlled trials
Methodological quality criteria
Randomized controlled
clinical trial: A clinical
study in which individual
participants are allocated
to intervention or control
groups by randomization
(intervention assigned by
researchers).
Key references: Higgins
and Green (2008);
Higgins et al. (2016);
Oxford Centre for
Evidence-based
Medicine (2016); Porta
et al. (2014)
2.1. Is randomization appropriately performed?
Explanations
In a randomized controlled trial, the allocation of a participant (or a data collection unit, e.g., a school) into the intervention or control group is based solely on chance.
Researchers should describe how the randomization schedule was generated. A simple statement such as ‘we randomly allocated’ or ‘using a randomized design’ is insufficient
to judge if randomization was appropriately performed. Also, assignment that is predictable such as using odd and even record numbers or dates is not appropriate. At minimum,
a simple allocation (or unrestricted allocation) should be performed by following a predetermined plan/sequence. It is usually achieved by referring to a published list of random
numbers, or to a list of random assignments generated by a computer. Also, restricted allocation can be performed such as blocked randomization (to ensure particular allocation
ratios to the intervention groups), stratified randomization (randomization performed separately within strata), or minimization (to make small groups closely similar with
respect to several characteristics). Another important characteristic to judge if randomization was appropriately performed is allocation concealment that protects assignment
sequence until allocation. Researchers and participants should be unaware of the assignment sequence up to the point of allocation. Several strategies can be used to ensure
allocation concealment such relying on a central randomization by a third party, or the use of sequentially numbered, opaque, sealed envelopes (Higgins et al., 2016).
2.2. Are the groups comparable at baseline?
Explanations
Baseline imbalance between groups suggests that there are problems with the randomization. Indicators from baseline imbalance include: “(1) unusually large differences
between intervention group sizes; (2) a substantial excess in statistically significant differences in baseline characteristics than would be expected by chance alone; (3) imbalance
in key prognostic factors (or baseline measures of outcome variables) that are unlikely to be due to chance; (4) excessive similarity in baseline characteristics that is not
compatible with chance; (5) surprising absence of one or more key characteristics that would be expected to be reported” (Higgins et al., 2016, p. 10).
2.3. Are there complete outcome data?
Explanations
Almost all the participants contributed to almost all measures. There is no absolute and standard cut-off value for acceptable complete outcome data. Agree among your team
what is considered complete outcome data in your field and apply this uniformly across all the included studies. For instance, in the literature, acceptable complete data value
ranged from 80% (Thomas et al., 2004; Zaza et al., 2000) to 95% (Higgins et al., 2016). Similarly, different acceptable withdrawal/dropouts rates have been suggested: 5% (de
Vet et al., 1997; MacLehose et al., 2000), 20% (Sindhu et al., 1997; Van Tulder et al., 2003) and 30% for a follow-up of more than one year (Viswanathan and Berkman, 2012).
2.4. Are outcome assessors blinded to the intervention provided?
Explanations
Outcome assessors should be unaware of who is receiving which interventions. The assessors can be the participants if using participant reported outcome (e.g., pain), the
intervention provider (e.g., clinical exam), or other persons not involved in the intervention (Higgins et al., 2016).
2.5 Did the participants adhere to the assigned intervention?
Explanations
To judge this criterion, consider the proportion of participants who continued with their assigned intervention throughout follow-up. “Lack of adherence includes imperfect
compliance, cessation of intervention, crossovers to the comparator intervention and switches to another active intervention.” (Higgins et al., 2016, p. 25).
5
3. Quantitative non-randomized studies Methodological quality criteria
Non-randomized studies are defined as any quantitative
studies estimating the effectiveness of an intervention or
studying other exposures that do not use randomization to
allocate units to comparison groups (Higgins and Green,
2008).
Common designs include (this list if not exhaustive):
Non-randomized controlled trials
The intervention is assigned by researchers, but there is no
randomization, e.g., a pseudo-randomization. A non-
random method of allocation is not reliable in producing
alone similar groups.
Cohort study
Subsets of a defined population are assessed as exposed,
not exposed, or exposed at different degrees to factors of
interest. Participants are followed over time to determine if
an outcome occurs (prospective longitudinal).
Case-control study
Cases, e.g., patients, associated with a certain outcome are
selected, alongside a corresponding group of controls.
Data is collected on whether cases and controls were
exposed to the factor under study (retrospective).
Cross-sectional analytic study
At one particular time, the relationship between health-
related characteristics (outcome) and other factors
(intervention/exposure) is examined. E.g., the frequency of
outcomes is compared in different population subgroups
according to the presence/absence (or level) of the
intervention/exposure.
Key references for non-randomized studies: Higgins and
Green (2008); Porta et al. (2014); Sterne et al. (2016);
Wells et al. (2000)
3.1. Are the participants representative of the target population?
Explanations
Indicators of representativeness include: clear description of the target population and of the sample (inclusion and exclusion criteria), reasons
why certain eligible individuals chose not to participate, and any attempts to achieve a sample of participants that represents the target
population.
3.2. Are measurements appropriate regarding both the outcome and intervention (or exposure)?
Explanations
Indicators of appropriate measurements include: the variables are clearly defined and accurately measured; the measurements are justified and
appropriate for answering the research question; the measurements reflect what they are supposed to measure; validated and reliability tested
measures of the intervention/exposure and outcome of interest are used, or variables are measured using ‘gold standard’.
3.3. Are there complete outcome data?
Explanations
Almost all the participants contributed to almost all measures. There is no absolute and standard cut-off value for acceptable complete outcome
data. Agree among your team what is considered complete outcome data in your field (and based on the targeted journal) and apply this
uniformly across all the included studies. For example, in the literature, acceptable complete data value ranged from 80% (Thomas et al., 2004;
Zaza et al., 2000) to 95% (Higgins et al., 2016). Similarly, different acceptable withdrawal/dropouts rates have been suggested: 5% (de Vet et
al., 1997; MacLehose et al., 2000), 20% (Sindhu et al., 1997; Van Tulder et al., 2003) and 30% for follow-up of more than one year
(Viswanathan and Berkman, 2012).
3.4. Are the confounders accounted for in the design and analysis?
Explanations
Confounders are factors that predict both the outcome of interest and the intervention received/exposure at baseline. They can distort the
interpretation of findings and need to be considered in the design and analysis of a non-randomized study. Confounding bias is low if there is
no confounding expected, or appropriate methods to control for confounders are used (such as stratification, regression, matching,
standardization, and inverse probability weighting).
3.5 During the study period, is the intervention administered (or exposure occurred) as intended?
Explanations
For intervention studies, consider whether the participants were treated in a way that is consistent with the planned intervention. Since the
intervention is assigned by researchers, consider whether there was a presence of contamination (e.g., the control group may be indirectly
exposed to the intervention) or whether unplanned co-interventions were present in one group (Sterne et al., 2016).
For observational studies, consider whether changes occurred in the exposure status among the participants. If yes, check if these changes are
likely to influence the outcome of interest, were adjusted for, or whether unplanned co-exposures were present in one group (Morgan et al.,
2017).
6
4. Quantitative descriptive studies Methodological quality criteria
Quantitative descriptive studies are “concerned with and
designed only to describe the existing distribution of
variables without much regard to causal relationships or
other hypotheses” (Porta et al., 2014, p. 72). They are used
to monitoring the population, planning, and generating
hypothesis (Grimes and Schulz, 2002).
Common designs include the following single-group
studies (this list if not exhaustive):
Incidence or prevalence study without comparison
group
In a defined population at one particular time, what is
happening in a population, e.g., frequencies of factors
(importance of problems), is described (portrayed).
Survey
“Research method by which information is gathered by
asking people questions on a specific topic and the data
collection procedure is standardized and well defined.”
(Bennett et al., 2011, p. 3).
Case series
A collection of individuals with similar characteristics are
used to describe an outcome.
Case report
An individual or a group with a unique/unusual outcome is
described in detail.
Key references: Critical Appraisal Skills Programme
(2017); Draugalis et al. (2008)
4.1. Is the sampling strategy relevant to address the research question?
Explanations
Sampling strategy refers to the way the sample was selected. There are two main categories of sampling strategies: probability sampling
(involve random selection) and non-probability sampling. Depending on the research question, probability sampling might be preferable. Non-
probability sampling does not provide equal chance of being selected. To judge this criterion, consider whether the source of sample is
relevant to the target population; a clear justification of the sample frame used is provided; or the sampling procedure is adequate.
4.2. Is the sample representative of the target population?
Explanations
There should be a match between respondents and the target population. Indicators of representativeness include: clear description of the target
population and of the sample (such as respective sizes and inclusion and exclusion criteria), reasons why certain eligible individuals chose not
to participate, and any attempts to achieve a sample of participants that represents the target population.
4.3. Are the measurements appropriate?
Explanations
Indicators of appropriate measurements include: the variables are clearly defined and accurately measured, the measurements are justified and
appropriate for answering the research question; the measurements reflect what they are supposed to measure; validated and reliability tested
measures of the outcome of interest are used, variables are measured using ‘gold standard’, or questionnaires are pre-tested prior to data
collection.
4.4. Is the risk of nonresponse bias low?
Explanations
Nonresponse bias consists of “an error of nonobservation reflecting an unsuccessful attempt to obtain the desired information from an eligible
unit.” (Federal Committee on Statistical Methodology, 2001, p. 6). To judge this criterion, consider whether the respondents and non-
respondents are different on the variable of interest. This information might not always be reported in a paper. Some indicators of low
nonresponse bias can be considered such as a low nonresponse rate, reasons for nonresponse (e.g., noncontacts vs. refusals), and statistical
compensation for nonresponse (e.g., imputation).
The nonresponse bias is might not be pertinent for case series and case report. This criterion could be adapted. For instance, complete data on
the cases might be important to consider in these designs.
4.5. Is the statistical analysis appropriate to answer the research question?
Explanations
The statistical analyses used should be clearly stated and justified in order to judge if they are appropriate for the design and research question,
and if any problems with data analysis limited the interpretation of the results.

7
5. Mixed methods studies Methodological quality criteria
Mixed methods (MM) research involves combining qualitative
(QUAL) and quantitative (QUAN) methods. In this tool, to be
considered MM, studies have to meet the following criteria (Creswell
and Plano Clark, 2017): (a) at least one QUAL method and one QUAN
method are combined; (b) each method is used rigorously in accordance
to the generally accepted criteria in the area (or tradition) of research
invoked; and (c) the combination of the methods is carried out at the
minimum through a MM design (defined a priori, or emerging) and the
integration of the QUAL and QUAN phases, results, and data.
Common designs include (this list if not exhaustive):
Convergent design
The QUAL and QUAN components are usually (but not necessarily)
concomitant. The purpose is to examine the same phenomenon by
interpreting QUAL and QUAN results (bringing data analysis together
at the interpretation stage), or by integrating QUAL and QUAN
datasets (e.g., data on same cases), or by transforming data (e.g.,
quantization of qualitative data).
Sequential explanatory design
Results of the phase 1 - QUAN component inform the phase 2 - QUAL
component. The purpose is to explain QUAN results using QUAL
findings. E.g., the QUAN results guide the selection of QUAL data
sources and data collection, and the QUAL findings contribute to the
interpretation of QUAN results.
Sequential exploratory design
Results of the phase 1 - QUAL component inform the phase 2 - QUAN
component. The purpose is to explore, develop and test an instrument
(or taxonomy), or a conceptual framework (or theoretical model). E.g.,
the QUAL findings inform the QUAN data collection, and the QUAN
results allow a statistical generalization of the QUAL findings.
Key references: Creswell et al. (2011); Creswell and Plano Clark,
(2017); O'Cathain (2010)
5.1. Is there an adequate rationale for using a mixed methods design to address the research question?
Explanations
The reasons for conducting a mixed methods study should be clearly explained. Several reasons can be invoked such as to
enhance or build upon qualitative findings with quantitative results and vice versa; to provide a comprehensive and complete
understanding of a phenomenon or to develop and test instruments (Bryman, 2006).
5.2. Are the different components of the study effectively integrated to answer the research question?
Explanations
Integration is a core component of mixed methods research and is defined as the “explicit interrelating of the quantitative and
qualitative component in a mixed methods study” (Plano Clark and Ivankova, 2015, p. 40). Look for information on how
qualitative and quantitative phases, results, and data were integrated (Pluye et al., 2018). For instance, how data gathered by both
research methods was brought together to form a complete picture (e.g., joint displays) and when integration occurred (e.g.,
during the data collection-analysis or/and during the interpretation of qualitative and quantitative results).
5.3. Are the outputs of the integration of qualitative and quantitative components adequately interpreted?
Explanations
This criterion is related to meta-inference, which is defined as the overall interpretations derived from integrating qualitative and
quantitative findings (Teddlie and Tashakkori, 2009). Meta-inference occurs during the interpretation of the findings from the
integration of the qualitative and quantitative components, and shows the added value of conducting a mixed methods study
rather than having two separate studies.
5.4. Are divergences and inconsistencies between quantitative and qualitative results adequately addressed?
Explanations
When integrating the findings from the qualitative and quantitative components, divergences and inconsistencies (also called
conflicts, contradictions, discordances, discrepancies, and dissonances) can be found. It is not sufficient to only report the
divergences; they need to be explained. Different strategies to address the divergences have been suggested such as reconciliation,
initiation, bracketing and exclusion (Pluye et al., 2009b). Rate this criterion ‘Yes’ if there is no divergence.
5.5. Do the different components of the study adhere to the quality criteria of each tradition of the methods involved?
Explanations
The quality of the qualitative and quantitative components should be individually appraised to ensure that no important threats to
trustworthiness are present. To appraise 5.5, use criteria for the qualitative component (1.1 to 1.5), and the appropriate criteria for
the quantitative component (2.1 to 2.5, or 3.1 to 3.5, or 4.1 to 4.5). The quality of both components should be high for the mixed
methods study to be considered of good quality. The premise is that the overall quality of a mixed methods study cannot exceed
the quality of its weakest component. For example, if the quantitative component is rated high quality and the qualitative
component is rated low quality, the overall rating for this criterion will be of low quality.
*Adapted from National Institute for Health Care Excellence. (2012). Methods for the development of nice public health guidance. London: National Institute for Health and Care Excellence; and Scottish Intercollegiate
Guidelines Network. (2017). Algorithm for classifying study design for questions of effectiveness. Retrieved December 1, 2017, from http://www.sign.ac.uk/assets/study_design.pdf.
8
Algorithm for selecting the study categories to rate in the MMAT*
.
9
References
Abbott, A. (1998). The causal devolution. Sociological Methods & Research, 27(2), 148-181.
Bennett, C., Khangura, S., Brehaut, J. C., Graham, I. D., Moher, D., Potter, B. K., et al. (2011). Reporting guidelines for survey research: An analysis of published guidance and reporting practices. PLoS
Medicine, 8(8), e1001069.
Bryman, A. (2006). Integrating quantitative and qualitative research: How is it done? Qualitative Research, 6(1), 97-113.
Creswell, J. W. (2013a). Qualitative inquiry and research design: Choosing among five approaches (3rd ed.). Thousand Oaks, CA: SAGE Publications.
Creswell, J. W. (2013b). Research design: Qualitative, quantitative, and mixed methods approaches. Thousand Oaks, CA: SAGE Publications.
Creswell, J. W., Klassen, A. C., Plano Clark, V. L., Smith, K. C. (2011). Best practices for mixed methods research in the health sciences. Bethesda, MD: Office of Behavioral and Social Sciences
Research, National Institutes of Health. http://obssr.od.nih.gov/mixed_methods_research.
Creswell, J. W., & Plano Clark, V. (2017). Designing and conducting mixed methods research (3rd ed.). Thousand Oaks, CA: SAGE Publications.
Critical Appraisal Skills Programme. (2017). CASP checklists. Retrieved December 1, 2017, from http://www.casp-uk.net/casp-tools-checklists.
de Vet, H. C., de Bie, R. A., van der Heijden, G. J., Verhagen, A. P., Sijpkes, P., & Knipschild, P. G. (1997). Systematic reviews on the basis of methodological criteria. Physiotherapy, 83(6), 284-289.
Draugalis, J. R., Coons, S. J., & Plaza, C. M. (2008). Best practices for survey research reports: A synopsis for authors and reviewers. American Journal of Pharmaceutical Education, 72(1),
Article 11.
Federal Committee on Statistical Methodology. (2001). Measuring and reporting sources of error in surveys. Washington DC: Statistical Policy Office, Office of Information and Regulatory Affairs,
Office of Management and Budget.
Grimes, D. A., & Schulz, K. F. (2002). Descriptive studies: What they can and cannot do. The Lancet, 359(9301), 145-149.
Higgins, J. P., & Green, S. (2008). Cochrane handbook for systematic reviews of interventions. Chichester, UK: Wiley Online Library.
Higgins, J. P. T., Sterne, J. A. C., Savović, J., Page, M. J., Hróbjartsson, A., Boutron, I., et al. (2016). A revised tool for assessing risk of bias in randomized trials. In Chandler, J., McKenzie, J., Boutron,
I. & Welch, V. (Eds.), Cochrane Methods. Cochrane Database of Systematic Reviews, Issue 10 (Suppl 1).
Hong, Q. N. (2018). Revision of the Mixed Methods Appraisal Tool (MMAT): A mixed methods study (Doctoral dissertation). Department of Family Medicine, McGill University, Montréal.
MacLehose, R. R., Reeves, B. C., Harvey, I. M., Sheldon, T. A., Russell, I. T., & Black, A. M. (2000). A systematic review of comparisons of effect sizes derived from randomised and non-randomised
studies. Health Technology Assessment, 4(34), 1-154.
Morgan, R., Sterne, J., Higgins, J., Thayer, K., Schunemann, H., Rooney, A., et al. (2017). A new instrument to assess Risk of Bias in Non-randomised Studies of Exposures (ROBINS-E): Application to
studies of environmental exposure. Abstracts of the Global Evidence Summit, Cape Town, South Africa. Cochrane Database of Systematic Reviews 2017, Issue 9 (Suppl 1).
https://doi.org/10.1002/14651858.CD201702.
O'Cathain, A. (2010). Assessing the quality of mixed methods research: Towards a comprehensive framework. In Tashakkori, A. & Teddlie, C. (Eds.), Handbook of Mixed methods in social and
behavioral research (pp. 531-555). Thousand Oaks, CA: SAGE Publications.
Oxford Centre for Evidence-based Medicine. (2016). Levels of evidence. Retrieved February 19, 2018, from https://www.cebm.net/2016/05/ocebm-levels-of-evidence/.
Pace, R., Pluye, P., Bartlett, G., Macaulay, A. C., Salsberg, J., Jagosh, J., et al. (2012). Testing the reliability and efficiency of the pilot Mixed Methods Appraisal Tool (MMAT) for systematic mixed
studies review. International Journal of Nursing Studies, 49(1), 47-53.
Plano Clark, V. L., & Ivankova, N. V. (2015). Mixed methods research: A guide to the field. Thousand Oaks, CA: SAGE Publications.
Pluye, P., Gagnon, M. P., Griffiths, F., Johnson-Lafleur, J. (2009a). A scoring system for appraising mixed methods research, and concomitantly appraising qualitative, quantitative and mixed methods
primary studies in mixed studies reviews. International Journal of Nursing Studies, 46(4), 529-546.
Pluye, P., Grad, R. M., Levine, A., & Nicolau, B. (2009b). Understanding divergence of quantitative and qualitative data (or results) in mixed methods studies. International Journal of Multiple Research
Approaches, 3(1), 58-72.
Pluye, P., Garcia Bengoechea, E., Granikov, V., Kaur, N., & Tang, D. L. (2018). A world of possibilities in mixed methods: Review of the combinations of strategies used to integrate the phases, results,
and qualitative and quantitative data. International Journal of Multiple Research Approaches, 10(1), 41-56.
Porta, M. S., Greenland, S., Hernán, M., dos Santos Silva, I., Last, J. M. (2014). A dictionary of epidemiology. New York: Oxford University Press.
Sandelowski, M. (2010). What's in a name? Qualitative description revisited. Research in Nursing and Health, 33(1), 77-84.
10
Schwandt, T. A. (2015). The SAGE dictionary of qualitative inquiry. Thousand Oaks, CA: SAGE Publications.
Sindhu, F., Carpenter, L., & Seers, K. (1997). Development of a tool to rate the quality assessment of randomized controlled trials using a Delphi technique. Journal of Advanced Nursing, 25(6), 1262-
1268.
Sterne, J. A., Hernán, M. A., Reeves, B. C., Savović, J., Berkman, N. D., Viswanathan, M., et al. (2016). ROBINS-I: A tool for assessing risk of bias in non-randomised studies of interventions. British
Medical Journal, 355(i4919).
Teddlie, C., & Tashakkori, A. (2009). Foundations of mixed methods research: Integrating quantitative and qualitative approaches in the social and behavioral sciences. Thousand Oaks, CA: SAGE
Publications.
Thomas, B. H., Ciliska, D., Dobbins, M., & Micucci, S. (2004). A process for systematically reviewing the literature: Providing the research evidence for public health nursing interventions. Worldviews
on Evidence-Based Nursing, 1(3), 176-184.
Van Tulder, M., Furlan, A., Bombardier, C., Bouter, L.,
& Editorial Board of the Cochrane Collaboration Back Review Group. (2003). Updated method guidelines for systematic reviews in the Cochrane
Collaboration Back Review Group. Spine, 28(12), 1290-1299.
Viswanathan, M., & Berkman, N. D. (2012). Development of the RTI item bank on risk of bias and precision of observational studies. Journal of Clinical Epidemiology, 65(2), 163-178.
Wells, G., Shea, B., O’connell, D., Peterson, J., Welch, V., Losos, M., et al. (2000). The Newcastle-Ottawa Scale (NOS) for assessing the quality of nonrandomised studies in meta-analyses. Retrieved
April 16, 2016, from http://www.ohri.ca/programs/clinical_epidemiology/nosgen.pdf.
Zaza, S., Wright-De Agüero, L. K., Briss, P. A., Truman, B. I., & Hopkins, D. P. (2000). Data collection instrument and procedure for systematic reviews in the guide to community preventive services.
American Journal of Preventive Medicine, 188(Suppl 1), 44-74.