KB

Week 6 EBP Critical Appraisal in Health Research

Benefits of Critical Appraisal

  • Health professionals need to make decisions based on the best available evidence.
  • Need to determine the quality of evidence to ensure it has a clinically significant impact.
  • The quality of a research study is related to how the research was conducted.
  • Findings should be a true and accurate representation, not due to biased methodology.

Peer Review

  • Check if a research study has been peer-reviewed.
  • Articles are reviewed by qualified researchers familiar with the topic before publication.
  • Many journals require several peer reviews before accepting an article.
  • Check the peer review process on the journal's homepage under author submission guidelines.
  • Peer review does not guarantee good quality research as the level of scrutiny varies.

Key Terms

  • Validity: Relates to research using a quantitative methodology.
  • Rigor: Relates to research using a qualitative methodology.

Applicability of Research

  • Consider whether the findings can be generalized to other settings.
    • In quantitative research, generalizability is referred to as external validity.
    • Control of bias is referred to as internal validity.
  • Researchers may prioritize internal validity (control of bias) at the expense of external validity (generalizability).
  • Ideally, research should be both good quality and able to be generalized.

Undertaking Critical Appraisal

  • Do not just read the abstract and discussion.
  • Critique each part of the reported research.
  • Read the entire article at a superficial level first, then read in detail and make notes.
  • Consider skipping the abstract initially and compare your understanding after detailed reading.

Non-Inferiority Trials

  • Designed to show that a treatment is equivalent to another treatment.
  • Relevant when a new treatment offers advantages like improved safety, convenience, compliance, or cost.
  • Example: Comparing telehealth delivery with in-clinic delivery.
  • Telehealth offers reduced costs and increased convenience but must demonstrate equivalent efficacy.

N-of-1 Trials

  • Also referred to as a single case study, it involves a single participant.
  • Random allocation can be used to determine the order of experimental and control interventions.
  • Used for people with rare health conditions; increasingly recognized for common conditions.

Critical Appraisal Tools

  • Guide you through questions evaluating research.
    • What was the clinical question?
    • Which study design was used and was it appropriate?
    • What were the characteristics of the sample and recruitment procedure?
    • What data were collected and how?
    • What was the independent variable, and how was it administered (for quantitative research)?
    • What other potential sources of bias may have affected the study?
    • What are the results, and are they applicable to practice?

Developing Appraisal Skills

  • With experience, you may not need a tool.
  • Recognize common quality pitfalls in specific areas of practice.
  • Critical appraisal tools ensure a thorough EBP (Evidence-Based Practice) process.

Validity

  • Concerns the integrity of the findings from a study.
  • Health practitioners must be aware of the main types of validity.

Internal Validity

  • Addresses causality: Can we draw a causal relationship between two variables?
  • Dependent variable (y): The variable researchers need to understand or explain why it varies.
  • Independent variable (x): The variable believed to produce variation in the dependent variable.
  • Need to be confident that x causes y and not an extraneous variable.
  • Example: Birth control classes and increased knowledge of birth control options.
  • Eliminate other plausible explanations to ensure the independent variable is responsible.

Post Hoc Fallacy:

  • The logical fallacy that because Y followed X, Y was caused by X.
  • In Latin: Post hoc ergo propter hoc or post hoc for short.
  • Other confounding factors might be responsible for changes detected.

External Validity

  • The degree to which findings can be generalized beyond the study participants.
  • Research is intended to guide practice, so generalizability is important.
  • Asks whether causal relationships can be generalized to different measures, persons, settings, and times.
  • Criticism: Highly controlled studies may only show treatment effectiveness in ideal circumstances.

Ecological Validity

  • Do the findings of a research study reflect real-life settings?
  • Researchers often focus on maximizing internal validity, potentially sacrificing external validity.
  • Academic journals and funding organizations often require high internal validity.

Efficacy Studies

  • Focus on demonstrating internal validity through highly controlled methodology.

Effectiveness Studies

  • Follow efficacy studies and are carried out in less controlled, real-life situations.

Establishing External Validity

  • Consider how well the research, using a sample, can be generalized to the population.
    • How many participants were involved, and was this a sufficient number to generalize to the population?
    • Was the sample size justified?
    • Is a clear description of the key characteristics of the sample provided? For example, the age and gender, and the time of onset of the health condition of interest.
    • If there were multiple groups in the study, were they comparable in terms of size and participant characteristics?
  • Were appropriate inclusion and exclusion criteria described?

Identifying Common Types of Bias

  • Bias can affect the internal and external validity of research.
  • Internal validity: Identifying a causal relationship between independent and dependent variables.
  • External validity: Factors like sample size and comparability of different groups.
  • Bias affects study results in a particular direction, favoring either the treatment or control group.

Categories of Bias

  • Sample or selection bias
  • Measurement or detection bias
  • Intervention or performance bias

Sample or Selection Bias

  • Volunteer or referral bias
  • Attention bias

Measurement or Detection Bias

  • Number of outcome measures used
  • Lack of ‘masked’ or ‘independent’ evaluation
  • Recall or memory bias

Intervention or Performance Bias

  • Contamination
  • Co-intervention
  • Timing of intervention
  • Site of intervention
  • Different administrators of the intervention

Measurement Errors

  • Categorized as random or systematic.
  • Random error: Unpredictable and cannot be controlled.
  • Systematic error: Predictable and can be identified and eliminated (e.g., imperfect calibration).

Types of Bias

Sample or Selection Bias

Volunteer or Referral Bias
  • Description: Participants volunteer for the study, potentially being more motivated.
  • Potential solution: Randomly select participants where possible; invite from waiting lists rather than advertising.
Attention Bias
  • Description: Awareness of the study's intention might cause participants to perform differently (placebo effect).
  • Potential solution: Include a control group and, if possible, a placebo treatment.

Intervention or Performance Bias

Contamination
  • Description: The control group inadvertently receives treatment.
  • Potential solution: Strict protocols for treatment delivery and control group management.
Co-intervention
  • Description: A participant receives another intervention simultaneously, influencing results.
  • Potential solution: Obtain information about medications or other interventions participants are undergoing.
Timing of Intervention
  • Description: Short duration may not allow noticeable change; long duration may lead to maturation.
  • Potential solution: Follow treatment protocols; use a control group to address maturation.
Site of Intervention
  • Description: The intervention site may affect the result.
  • Potential solution: Ensure consistency across all treatment sites.
Different administrators of the intervention
  • Description: Different therapists deliver the treatment, leading to variability.
  • Potential solution: Address therapist variability to ensure more consistent results

Other Limitations in Quantitative Research

  • Sample characteristics
  • Dropouts
  • Method and frequency of measurement

Sample

  • Questions to ask:
    • How many participants were involved, and was this a sufficient number to be able to generalize the results of population?
    • Was the sample size justified? Preliminary studies may involve small sample sizes, but if the researchers aim to demonstrate efficacy, then a larger sample is needed.
    • Was a clear description of the key characteristics of the same provided (e.g. the age and gender of participants, and the onset of the health condition of interest)?
    • If there were multiple groups in the study, were they comparable in terms of size and participant characteristics?
  • Were appropriate inclusion and exclusion criteria described?

Dropouts

  • Also known as experimental mortality or attrition.
  • Questions to ask:
    • Were the number of drops outs reported?
    • Were the reasons for the dropouts documented?
    • How did the researchers manage the analysis of the data to deal with any missing data caused by participants dropping out?

Measurement

  • Questions to ask:
    • How frequently were the outcomes measured; for example, were they measured before and after treatment, or were short-term and long-term follow-up data also collected?
    • Did the researchers report whether the outcomes measures used are well- established as being reliable and valid?