Clinical Research Methodology 3: Randomized Controlled Trials
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
From Explanatory to Pragmatic Clinical Trials: a New Era for Effectiveness Research
From Explanatory to Pragmatic Clinical Trials: A New Era for Effectiveness Research Jerry Jarvik, M.D., M.P.H. Professor of Radiology, Neurological Surgery and Health Services Adjunct Professor Orthopedic Surgery & Sports Medicine and Pharmacy Director, Comparative Effectiveness, Cost and Outcomes Research Center (CECORC) UTSW April 2018 Acknowledgements •NIH: UH2 AT007766; UH3 AT007766 •NIH P30AR072572 •PCORI: CE-12-11-4469 Disclosures Physiosonix (ultrasound company): Founder/stockholder UpToDate: Section Editor Evidence-Based Neuroimaging Diagnosis and Treatment (Springer): Co-Editor The Big Picture Comparative Effectiveness Evidence Based Practice Health Policy Learning Healthcare System So we need to generate evidence Challenge #1: Clinical research is slow • Tradi>onal RCTs are slow and expensive— and rarely produce findings that are easily put into prac>ce. • In fact, it takes an average of 17 ye ars before research findings Howlead pragmaticto widespread clinical trials canchanges improve in care. practice & policy Challenge #1: Clinical research is slow “…rarely produce findings that are easily put into prac>ce.” Efficacy vs. Effectiveness Efficacy vs. Effectiveness • Efficacy: can it work under ideal conditions • Effectiveness: does it work under real-world conditions Challenge #2: Clinical research is not relevant to practice • Tradi>onal RCTs study efficacy “If we want of txs for carefully selected more evidence- popula>ons under ideal based pr actice, condi>ons. we need more • Difficult to translate to real practice-based evidence.” world. Green, LW. American Journal • When implemented into of Public Health, 2006. How pragmatic clinical trials everyday clinical prac>ce, oZen seecan a “ voltage improve drop practice”— drama>c & decreasepolicy from efficacy to effec>veness. -
Evolution of Clinical Trials Throughout History
Evolution of clinical trials throughout history Emma M. Nellhaus1, Todd H. Davies, PhD1 Author Affiliations: 1. Office of Research and Graduate Education, Marshall University Joan C. Edwards School of Medicine, Huntington, West Virginia The authors have no financial disclosures to declare and no conflicts of interest to report. Corresponding Author: Todd H. Davies, PhD Director of Research Development and Translation Marshall University Joan C. Edwards School of Medicine Huntington, West Virginia Email: [email protected] Abstract The history of clinical research accounts for the high ethical, scientific, and regulatory standards represented in current practice. In this review, we aim to describe the advances that grew from failures and provide a comprehensive view of how the current gold standard of clinical practice was born. This discussion of the evolution of clinical trials considers the length of time and efforts that were made in order to designate the primary objective, which is providing improved care for our patients. A gradual, historic progression of scientific methods such as comparison of interventions, randomization, blinding, and placebos in clinical trials demonstrates how these techniques are collectively responsible for a continuous advancement of clinical care. Developments over the years have been ethical as well as clinical. The Belmont Report, which many investigators lack appreciation for due to time constraints, represents the pinnacle of ethical standards and was developed due to significant misconduct. Understanding the history of clinical research may help investigators value the responsibility of conducting human subjects’ research. Keywords Clinical Trials, Clinical Research, History, Belmont Report In modern medicine, the clinical trial is the gold standard and most dominant form of clinical research. -
U.S. Investments in Medical and Health Research and Development 2013 - 2017 Advocacy Has Helped Bring About Five Years of Much-Needed Growth in U.S
Fall 2018 U.S. Investments in Medical and Health Research and Development 2013 - 2017 Advocacy has helped bring about five years of much-needed growth in U.S. medical and health research investment. More advocacy is critical now to ensure our nation steps up in response to health threats that we can—we must—overcome. More Than Half Favor Doubling Federal Spending on Medical Research Do you favor or oppose doubling federal spending on medical research over the next five years? 19% Not Sure 23% 8% Strongly favor Strongly oppose 16% 35% Somewhat oppose Somewhat favor Source: A Research!America survey of U.S. adults conducted in partnership with Zogby Analytics in January 2018. Research!America 3 Introduction Investment1 in medical and health research and development (R&D) in the U.S. grew by $38.8 billion or 27% from 2013 to 2017. Industry continues to invest more than any other sector, accounting for 67% of total spending in 2017, followed by the federal government at 22%. Federal investments increased from 2016 to 2017, the second year of growth after a dip from 2014 to 2015. Overall, federal investment increased by $6.1 billion or 18.4% from 2013 to 2017, but growth has been uneven across federal health agencies. Investment by other sectors, including academic and research institutions, foundations, state and local governments, and voluntary health associations and professional societies, also increased from 2013 to 2017. Looking ahead, medical and health R&D spending is expected to move in an upward trajectory in 2018 but will continue to fall short relative to the health and economic impact of major health threats. -
Using Randomized Evaluations to Improve the Efficiency of US Healthcare Delivery
Using Randomized Evaluations to Improve the Efficiency of US Healthcare Delivery Amy Finkelstein Sarah Taubman MIT, J-PAL North America, and NBER MIT, J-PAL North America February 2015 Abstract: Randomized evaluations of interventions that may improve the efficiency of US healthcare delivery are unfortunately rare. Across top journals in medicine, health services research, and economics, less than 20 percent of studies of interventions in US healthcare delivery are randomized. By contrast, about 80 percent of studies of medical interventions are randomized. The tide may be turning toward making randomized evaluations closer to the norm in healthcare delivery, as the relevant actors have an increasing need for rigorous evidence of what interventions work and why. At the same time, the increasing availability of administrative data that enables high-quality, low-cost RCTs and of new sources of funding that support RCTs in healthcare delivery make them easier to conduct. We suggest a number of design choices that can enhance the feasibility and impact of RCTs on US healthcare delivery including: greater reliance on existing administrative data; measuring a wide range of outcomes on healthcare costs, health, and broader economic measures; and designing experiments in a way that sheds light on underlying mechanisms. Finally, we discuss some of the more common concerns raised about the feasibility and desirability of healthcare RCTs and when and how these can be overcome. _____________________________ We are grateful to Innessa Colaiacovo, Lizi Chen, and Belinda Tang for outstanding research assistance, to Katherine Baicker, Mary Ann Bates, Kelly Bidwell, Joe Doyle, Mireille Jacobson, Larry Katz, Adam Sacarny, Jesse Shapiro, and Annetta Zhou for helpful comments, and to the Laura and John Arnold Foundation for financial support. -
Certification Accredited Schools List
= Academic Clinical Research Programs Which Currently Meet Waiver Requirements for the Academy of Clinical Research Professional (Updated October 15, 2015) This list includes academic clinical research programs which, as of the above date, have been evaluated and found to meet the current waiver requirements established by the Academy of Clinical Research Professionals (Academy) Board of Trustees. Programs that meet the waiver requirements allow a candidate to waive 1,500 hours of hands-on experience performing the Essential Duties of the program to which the candidate has made application. This list is not to be considered exhaustive and is not designed to represent all the academic offerings available. Appearance on this list is not to be construed as an endorsement of its content or format by the Academy or a guarantee that a candidate will be eligible for certification. All individuals interested in enrolling in a program of study should evaluate any program on the basis of his or her personal needs. (* - programs marked with an asterisk MAY be accepted but acceptance will be dependent on the complete listing of courses completed) Individuals seeking additional information about each program should contact the institution directly. Academic Institution Program Name ( Click Title for Web Navigation) Academy of Applied Pharmaceutical Sciences Clinical Research Diploma Program American University of Health Sciences Master of Science in Clinical Research Anoka-Ramsey Community College Clinical Research Professional - Certificate Graduate -
Clinical Research Services
Clinical Research Services Conducting efficient, innovative clinical research from initial planning to closeout WHY LEIDOS LIFE SCIENCES? In addition to developing large-scale technology programs for U.S. federal f Leidos has expertise agencies with a focus on health, Leidos provides a broad range of clinical supporting all product services and solutions to researchers, product sponsors, U.S. government development phases agencies, and other biomedical enterprises, hospitals, and health systems. for vaccines, drugs, and Our broad research experience enables us to support programs throughout biotherapeutics the development life cycle: from concept through the exploratory, f Leidos understands the development, pre-clinical, and clinical phases, as well as with product need to control clinical manufacturing and launching. Leidos designs and develops customized project scope, timelines, solutions that support groundbreaking medical research, optimize business and cost while remaining operations, and expedite the discovery of safe and effective medical flexible amidst shifting products. We apply our technical knowledge and experience in selecting research requirements institutions, clinical research organizations, and independent research f Leidos remains vendor- facilities to meet the specific needs of each clinical study. We also manage neutral while fostering and team integration, communication, and contracts throughout the project. managing collaborations Finally, Leidos has a proven track record of ensuring that research involving with -
Placebo-Controlled Trials of New Drugs: Ethical Considerations
Reviews/Commentaries/Position Statements COMMENTARY Placebo-Controlled Trials of New Drugs: Ethical Considerations DAVID ORENTLICHER, MD, JD als, placebo controls are not appropriate when patients’ health would be placed at significant risk (8–10). A placebo- controlled study for a new hair- uch controversy exists regarding sufficiently more effective than placebo to thickening agent could be justified; a the ethics of placebo-controlled justify its use. Finally, not all established placebo-controlled study for patients M trials in which an experimental therapies have been shown to be superior with moderate or severe hypertension therapy will compete with an already es- to placebo. If newer drugs are compared would not be acceptable (11). Similarly, if tablished treatment (or treatments). In with the unproven existing therapies, an illness causes problems when it goes such cases, argue critics, patients in the then patients may continue to receive untreated for a long period of time, a 52- control arm of the study should receive an drugs that are harmful without being week study with a placebo control is accepted therapy rather than a placebo. helpful. much more difficult to justify than a By using an active and effective drug, the Moreover, say proponents of placebo 6-week study (12). control patients would not be placed at controls, patients can be protected from When David S.H. Bell (13) explains risk for deterioration of their disease, and harm by “escape” criteria, which call for why placebo controls are unacceptable the study would generate more meaning- withdrawal from the trial if the patient for new drugs to treat type 2 diabetes, he ful results for physicians. -
Interpreting Indirect Treatment Comparisons and Network Meta
VALUE IN HEALTH 14 (2011) 417–428 available at www.sciencedirect.com journal homepage: www.elsevier.com/locate/jval SCIENTIFIC REPORT Interpreting Indirect Treatment Comparisons and Network Meta-Analysis for Health-Care Decision Making: Report of the ISPOR Task Force on Indirect Treatment Comparisons Good Research Practices: Part 1 Jeroen P. Jansen, PhD1,*, Rachael Fleurence, PhD2, Beth Devine, PharmD, MBA, PhD3, Robbin Itzler, PhD4, Annabel Barrett, BSc5, Neil Hawkins, PhD6, Karen Lee, MA7, Cornelis Boersma, PhD, MSc8, Lieven Annemans, PhD9, Joseph C. Cappelleri, PhD, MPH10 1Mapi Values, Boston, MA, USA; 2Oxford Outcomes, Bethesda, MD, USA; 3Pharmaceutical Outcomes Research and Policy Program, School of Pharmacy, School of Medicine, University of Washington, Seattle, WA, USA; 4Merck Research Laboratories, North Wales, PA, USA; 5Eli Lilly and Company Ltd., Windlesham, Surrey, UK; 6Oxford Outcomes Ltd., Oxford, UK; 7Canadian Agency for Drugs and Technologies in Health (CADTH), Ottawa, ON, Canada; 8University of Groningen / HECTA, Groningen, The Netherlands; 9University of Ghent, Ghent, Belgium; 10Pfizer Inc., New London, CT, USA ABSTRACT Evidence-based health-care decision making requires comparisons of all of randomized, controlled trials allow multiple treatment comparisons of relevant competing interventions. In the absence of randomized, con- competing interventions. Next, an introduction to the synthesis of the avail- trolled trials involving a direct comparison of all treatments of interest, able evidence with a focus on terminology, assumptions, validity, and statis- indirect treatment comparisons and network meta-analysis provide use- tical methods is provided, followed by advice on critically reviewing and in- ful evidence for judiciously selecting the best choice(s) of treatment. terpreting an indirect treatment comparison or network meta-analysis to Mixed treatment comparisons, a special case of network meta-analysis, inform decision making. -
Generating Real-World Evidence by Strengthening Real-World Data Sources
Generating Real-World Evidence by Strengthening Real-World Data Sources Using “real-world evidence” to bring new “Every day, health care professionals are updating patients’ treatments to patients as part of the 21st electronic health records with data on clinical outcomes Century Cures Act (the “Cures Act”) is a key resulting from medical interventions used in routine clinical priority for the Department of Health and practice. As our experience with new medical products Human Services (HHS). Specifically, the Cures expands, our knowledge about how to best maximize their Act places focus on the use of real-world data benefits and minimize potential risks sharpens with each data to support regulatory decision-making, including point we gather. Every clinical use of a product produces data the approval of new indications for existing that can help better inform us about its safety and efficacy.” drugs in order to make drug development faster jacqueline corrigan-curay, md, jd and more efficient. As such, the Cures Act director of the office of medical policy in fda’s center for drug evaluation and research has tasked the Food and Drug Administration (FDA) to develop a framework and guidance for evaluating real-world evidence in the context of drug regulation.1 Under the FDA’s framework, real-world evidence (RWE) is generated by different study designs or analyses, including but not limited to, randomized trials like large simple trials, pragmatic trials, and observational studies. RWE is the clinical evidence about the use, potential benefits, and potential risks of a medical product based on an analysis of real-world data (RWD). -
Patient Reported Outcomes (PROS) in Performance Measurement
Patient-Reported Outcomes Table of Contents Introduction ............................................................................................................................................ 2 Defining Patient-Reported Outcomes ............................................................................................ 2 How Are PROs Used? ............................................................................................................................ 3 Measuring research study endpoints ........................................................................................................ 3 Monitoring adverse events in clinical research ..................................................................................... 3 Monitoring symptoms, patient satisfaction, and health care performance ................................ 3 Example from the NIH Collaboratory ................................................................................................................... 4 Measuring PROs: Instruments, Item Banks, and Devices ....................................................... 5 PRO Instruments ............................................................................................................................................... 5 Item Banks........................................................................................................................................................... 8 Devices ................................................................................................................................................................. -
A National Strategy to Develop Pragmatic Clinical Trials Infrastructure
A National Strategy to Develop Pragmatic Clinical Trials Infrastructure Thomas W. Concannon, Ph.D.1,2, Jeanne-Marie Guise, M.D., M.P.H.3, Rowena J. Dolor, M.D., M.H.S.4, Paul Meissner, M.S.P.H.5, Sean Tunis, M.D., M.P.H.6, Jerry A. Krishnan, M.D., Ph.D.7,8, Wilson D. Pace, M.D.9, Joel Saltz, M.D., Ph.D.10, William R. Hersh, M.D.3, Lloyd Michener, M.D.4, and Timothy S. Carey, M.D., M.P.H.11 Abstract An important challenge in comparative effectiveness research is the lack of infrastructure to support pragmatic clinical trials, which com- pare interventions in usual practice settings and subjects. These trials present challenges that differ from those of classical efficacy trials, which are conducted under ideal circumstances, in patients selected for their suitability, and with highly controlled protocols. In 2012, we launched a 1-year learning network to identify high-priority pragmatic clinical trials and to deploy research infrastructure through the NIH Clinical and Translational Science Awards Consortium that could be used to launch and sustain them. The network and infrastructure were initiated as a learning ground and shared resource for investigators and communities interested in developing pragmatic clinical trials. We followed a three-stage process of developing the network, prioritizing proposed trials, and implementing learning exercises that culminated in a 1-day network meeting at the end of the year. The year-long project resulted in five recommendations related to developing the network, enhancing community engagement, addressing regulatory challenges, advancing information technology, and developing research methods. -
Adaptive Designs in Clinical Trials: Why Use Them, and How to Run and Report Them Philip Pallmann1* , Alun W
Pallmann et al. BMC Medicine (2018) 16:29 https://doi.org/10.1186/s12916-018-1017-7 CORRESPONDENCE Open Access Adaptive designs in clinical trials: why use them, and how to run and report them Philip Pallmann1* , Alun W. Bedding2, Babak Choodari-Oskooei3, Munyaradzi Dimairo4,LauraFlight5, Lisa V. Hampson1,6, Jane Holmes7, Adrian P. Mander8, Lang’o Odondi7, Matthew R. Sydes3,SofíaS.Villar8, James M. S. Wason8,9, Christopher J. Weir10, Graham M. Wheeler8,11, Christina Yap12 and Thomas Jaki1 Abstract Adaptive designs can make clinical trials more flexible by utilising results accumulating in the trial to modify the trial’s course in accordance with pre-specified rules. Trials with an adaptive design are often more efficient, informative and ethical than trials with a traditional fixed design since they often make better use of resources such as time and money, and might require fewer participants. Adaptive designs can be applied across all phases of clinical research, from early-phase dose escalation to confirmatory trials. The pace of the uptake of adaptive designs in clinical research, however, has remained well behind that of the statistical literature introducing new methods and highlighting their potential advantages. We speculate that one factor contributing to this is that the full range of adaptations available to trial designs, as well as their goals, advantages and limitations, remains unfamiliar to many parts of the clinical community. Additionally, the term adaptive design has been misleadingly used as an all-encompassing label to refer to certain methods that could be deemed controversial or that have been inadequately implemented. We believe that even if the planning and analysis of a trial is undertaken by an expert statistician, it is essential that the investigators understand the implications of using an adaptive design, for example, what the practical challenges are, what can (and cannot) be inferred from the results of such a trial, and how to report and communicate the results.