
This article is part of the NiaHealth Research Process series.
Evidence-based practice, also known as evidence-based medicine and evidence-based health care, involves integrating the best research evidence with clinical expertise and patients' circumstances and values (Sackett et al., 1996). This definition highlights the importance of utilizing research evidence with clinical expertise and patient characteristics and preferences, accounting for the notion that evidence cannot necessarily be applied generally to all patients, often requiring individualization.
The research side of evidence-based practice focuses on the development of the identification and synthesis of the highest quality evidence. There is a general consensus on what constitutes high quality evidence, although there are some exceptions. This hierarchy is typically conceptualized as an evidence pyramid where the highest level of evidence is at the top of the pyramid, with the validity of the evidence decreasing at lower levels, although sometimes research at a lower level of the pyramid will be more appropriate to answer a specific research question (e.g. the design is most appropriate or the quality is better than studies that are higher on the pyramid). The best choice of study may also vary based on clinical context. For example, a study done in a certain population may be more applicable to an individual from that population than a meta-analysis that did not stratify between populations.

Types of studies and their respective strengths and weaknesses are summarized below (also reviewed in detail in Wallace et al., 2022).
When conducted in a rigorous manner, systematic reviews and meta-analysis are generally considered to be the highest quality study designs. Systematic reviews synthesize available evidence evaluated from multiple distinct studies and often include meta-analyses, which combine numerical outcomes as if they were the result of a single larger study. These methodological approaches require a rigorous literature search of published and unpublished literature, appraisal of each relevant article for bias, and a quality assessment of the level of evidence for each outcome.
Systematic reviews and meta-analyses make readers aware of the current state of the evidence and gaps where further studies are needed. However, they do have limitations. Inclusion of studies with biased results can result in an inaccurate qualitative and/or quantitative synthesis (garbage in to garbage out) and when heterogeneous studies are synthesized quantitatively, taking the results at face value can lead to biased conclusions. It is common for meta-analyses conducted by different research groups to draw different conclusions based on study inclusion criteria.
RCTs are prospective studies that evaluate the effectiveness of an intervention through comparison with placebo or the existing standard of care. They provide higher internal validity than can typically be achieved with observational methods, and RCTs are the only type of study definitively able to establish causation rather than correlation or association. Randomization equalizes the distribution of confounding factors in each group, while allocation concealment reduces selection bias and blinding can help to address observation bias. RCTs may have limited generalizability, particularly if they have numerous/restrictive exclusion criteria or low consent rates. Additionally, attrition bias can occur if loss to follow-up or drop-out rates are unbalanced between arms.
Cohort studies identify a specific population in which a subset of individuals has experienced a particular exposure and compare the rates of disease development in exposed to unexposed individuals over time; they can be either retrospective or prospective. Cohort studies are a good design when the outcome of interest is common, and they also have the benefit of being able to measure multiple outcomes. Although observational studies cannot confirm causation, evaluating the time course of exposures and outcomes permits generation of hypotheses and can build evidence towards causality. Cohort studies are prone to confounding bias due to their observational design. Even when measured confounders are appropriately addressed statistically, not all confounders are measured or appreciated and may therefore be unaccounted for. Cohort studies may also suffer from attrition bias due to loss to follow-up, as well as surveillance bias, because certain diseases may be more likely to be detected in study participants than in the general population due to increased monitoring.
In cross-sectional studies, population data are collected at a specific point in time with simultaneous measurement of the exposure and outcome. These studies are often used to evaluate diagnostic accuracy of a test or measure. They cannot ascertain or suggest causality as the exposure and outcome occur at a single point in time.
Case-control studies compare groups of individuals with and without a specific disease with the aim of identifying disease risk factors. They are useful for studying rare diseases which would otherwise require very large cohorts to evaluate. Selection bias can occur in case-control studies if the control group is selected from a different population than cases, and recall bias may occur if cases are more likely to remember an exposure than controls or vice versa.
These studies provide a description of an individual clinical case or a series of related cases. They offer an opportunity for clinicians to share their experience caring for patients with unusual presentations or rare diseases. However, drawing conclusions from these study types to answer clinical questions is not advised, due to potential biases resulting from the small number of patients described and the lack of comparison groups.
Basic science and translational studies, including in vitro and animal studies, often serve as the foundation for clinical research. This type of research is typically used to drive the formulation of mechanistic hypotheses and ascertain the safety of drugs and devices before further study in humans. However, the results of these studies are not appropriate for clinical decision-making.
A second pyramid, aimed at clinical decision-makers, has also been proposed (DiCenso et al., 2009). This 6S Pyramid leverages resources that are available to facilitate rapid synthesis of high quality research.
The bottom tiers of the 6S Pyramid incorporate the study designs shown in the Evidence Pyramid and, above those, the systematic reviews and meta-analyses (labeled Syntheses). The higher tiers of evidence include research that has been further aggregated and summarized by experts.
.webp)
Systems (also known as Clinical Decision Support Systems) involve the integration of information from the lower levels of the evidence hierarchy with individual patient records. These are felt to represent the ideal source of evidence for clinical decision-making as they provide guidance that is both evidence-based and personalized.
These are regularly updated clinical guidelines or textbooks that integrate evidence-based information about specific clinical problems (e.g. Canadian Guidelines, UpToDate).
Synopses of syntheses summarize information found in systematic reviews (e.g. Health and Technology Assessment Database, Cochrane Summaries of Cochrane systematic reviews).
Syntheses include systematic reviews and meta-analyses (e.g. Cochrane Library, PubMed Health).
These resources provide summaries of evidence from individual high-quality studies (e.g. Tools for Practice and Evidence-based medicine).
These are unique research studies conducted to answer specific clinical questions. Single studies (as well as systematic reviews and meta-analyses) can be identified through searching databases such as OVID, PubMed, Web of Science, and Scopus.
This includes any literature that has not been published through traditional means. Grey literature includes conference abstracts, pre-prints, white papers, and sometimes, lay media sources. It is often excluded from large databases of traditionally published literature and requires specific search strategies to identify.
The Grading of Recommendations, Assessment, Development, and Evaluation (GRADE) tool was established for assessing the quality of evidence for interventions (Prasad, 2024). There are different versions of the tool for prognostic factors (Foroutan et al., 2020) and diagnostic tests (Schunemann et al., 2019).
Although the quality of evidence is a continuum, the GRADE approach results in classification of the quality of a body of evidence into one of the following four grades:
The GRADE handbook notes that the low and very low categories can be combined where appropriate.
The following steps are used to assess strength of evidence:
NiaHealth uses aspects of the GRADE approach to determine the quality and certainty of the evidence, similar to the approach described in the Editorial Policy of the medical database UpToDate.
Note that, for evidence to support a diagnostic test or prognostic factor, RCTs may not always be the most appropriate or available study design, and well-designed observational studies may provide enough evidence for a high quality grade.
If starting with RCTs, start by assuming high quality, but then check for:
Grade down from high to moderate or low depending on limitations.
If no RCTs (including indirectly relevant trials), start by assuming low quality, but then check for:
Grade up to moderate or even high depending on special strengths.
Our editorial standards & process
At NiaHealth, our mission is to make proactive health possible for all Canadians—by combining science with humanity. We believe that rigorous, evidence-informed health information should never feel out of reach. Every word we publish is intentional. We choose language that empowers rather than overwhelms, clarifies rather than complicates, and respects the lived experiences behind every health question. Learn more here.