Clinical-Study-Design-Virl.Pdf

Clinical-Study-Design-Virl.Pdf

Clinical Study Design Lecturer: Dr. Bryant England Purpose: Give an introduction to clinical study designs to improve ability to critically appraise literature and to encourage/make reading methods easier! Theory • Hill’s criteria for causation (strength, consistency, specificity, temporality, biological gradient, biological plausibility, coherence, experimental evidence, and analogy) • What makes a good study? Ask important questions and obtain valid results Bias: • Chance ( random error – statistics problem) vs Bias ( systematic error = design problem) • Bias in studies makes me SIC – Selection, Information, Confounding • Selection Bias – How are we selecting people for our study or selecting who stayed in our study o “Relationship between exposure and disease is different for those who participate” – Selection into a study & surveillance/selection out of study (retention) • Confounding Bias – Is there another variable that explains our association? o Risk factor for the disease o Associated with the exposure o Must not be affected by exposure (not an intermediate) • Information Bias – How are we measuring things? o “Bias in estimating an effect caused by measurement errors” o Special type of Information Bias – Misclassification – Measurement error of discrete variables (categorical data) Take Home: There are a lot of ways to screw up in research – we need to be on the lookout for them Application o Randomized Controlled Trial – we are testing whether randomization leads to better outcomes. • Define what is the population you want to study in • Randomize – Experimental therapy vs Control therapy • Study endpoint: Outcome assessment (study follow-up) o Randomization – chance determines intervention or control so that balances known and unknown confounders • Eliminates selection & confounding bias • Main Types: Simple, Block, Stratified o Outcomes – we want the appropriate study outcomes • Before you start the study, decide what is the most clinically relevant and validated outcome in the study (primary outcome) • All other relevant measures (secondary outcome) • Clinical Trial Statistics o Typical stats: incidence rates, relative risk, absolute risk difference, hazard ratio, p values o Effect size = magnitude of difference in outcomes between groups o P value = hypothesis test (Smaller P value does NOT equal stronger effect) o Power calculations – most important to critique when study findings are null o Pearl: First sentence of the method has a lot of information on RCT description o Other study designs when a RCT is not possible: Cohort, Case-Control, Case Study, Ecological, Systematic Review & Meta-Analysis, Basic Science designs Pearl: Design (and critique) the study like a RCT even if it is not a RCT Q&A Pearls: o In general, you do not want to change the primary outcome of the study after designing and beginning the study. However, if by the time a study has been initiated a more clinically relevant outcome with better psychometric properties has been developed/validation, it may make sense to change. Changing the primary outcome should never be because you don’t like what your data is showing! .

View Full Text

Details

  • File Type
    pdf
  • Upload Time
    -
  • Content Languages
    English
  • Upload User
    Anonymous/Not logged-in
  • File Pages
    2 Page
  • File Size
    -

Download

Channel Download Status
Express Download Enable

Copyright

We respect the copyrights and intellectual property rights of all users. All uploaded documents are either original works of the uploader or authorized works of the rightful owners.

  • Not to be reproduced or distributed without explicit permission.
  • Not used for commercial purposes outside of approved use cases.
  • Not used to infringe on the rights of the original creators.
  • If you believe any content infringes your copyright, please contact us immediately.

Support

For help with questions, suggestions, or problems, please contact us