Oseltamivir: the Real World Data

Total Page:16

File Type:pdf, Size:1020Kb

Oseltamivir: the Real World Data ANALYSIS Oseltamivir: the real world data To coincide with the publication of an updated systematic review by the Cochrane group the BMJ invited Nick Freemantle and colleagues to consider the current status of observational studies of oseltamivir and their influence on policy and practice seltamivir has become a mainstay in factors or confounders and are at the mercy of the treatment of influenza, although how accurately these factors are recorded. For there is no evidence from clini- example clinicians use their judgment to decide cal trials that it reduces mortality.1 which treatment participants should receive, and In 2009 we published a review of this will be influenced by whom they recognise to Oobservational studies examining the effects of be at higher risk. Their judgment therefore forms oseltamivir in influenza. It was based on a list unrecorded information on risk that cannot be of studies provided by Roche, the manufacturer, captured in a statistical model,4 which gives rise that it claimed provided evidence of the drug’s to what is known as confounding by indication. “real life” value.2 The evidence supported a role This confounding is relevant to observational for oseltamivir in reducing pneumonia and other data on influenza patients because doctors may complications of influenza in otherwise healthy be more likely to treat patients who are sicker. adults. But we found no evidence that it was Similarly, treatment may be given only to associated with mortality, little information those who survive long enough to receive on safety, and none on pregnancy related it, or remained ill for longer, termed survi- MALCOLM WILLETT MALCOLM outcomes. We have now updated our evidence vorship bias. by doing a systematic search of published obser- example, because of a small P value) then the Survivorship bias may play out in different vational studies. Here we explain the advantages experimental treatment must be the explana- ways in observational studies of oseltamivir. Jain and drawbacks of observational data and what tion. However, even impeccably designed ran- and colleagues describe a multivariable analy- they tell us about oseltamivir. domised controlled trials have limitations. For sis where receiving antiviral therapy within two Worldwide influenza is a major public health example, trials often include participants who days after the onset of illness was associated with concern. During an epidemic about 5-15% of are relatively healthy, young, male, lacking reduced mortality, but their analysis was open the global population develop an upper res- comorbidities, or more motivated and adher- to survivorship bias and confounding by indi- piratory tract infection, 3 to 5 million people ent to treatment than patients seen in clinical cation.5 It is likely that a patient who presented contract severe disease, and there are between practice, so the results may be not be able to be atypically and did not received early antivirals 250 000 and 500 000 influenza related deaths.3 generalised to a wider population. Furthermore, but recovered well might be spared oseltamivir Vaccination is the leading public health con- trials of relatively short duration in selected par- and perhaps not included in the study or data trol measure to prevent disease and reduce the ticipants without other risk factors may not pro- set. On the other hand a sicker or deteriorating effect of epidemics, but neuraminidase inhibi- vide evidence of real world safety. patient might be treated even though late treat- tors (oseltamivir or zanamivir) or M2 ion channel ment was not recommended. When a sicker blockers (amantadine or rimatidine) may be pre- What do observational data have to offer? group of patients is treated late, or when subjects scribed to reduce the duration of symptoms and Observational studies have some advantages who die early (before the opportunity to receive improve clinical outcomes. The use of neurami- over randomised trials. They can be conducted antiviral treatment) are included in an analysis nidase inhibitors increased substantially during in actual populations receiving treatment and as “untreated,” survivorship bias will make the 2009 pandemic of influenza A/H1N1, partly can be conducted rapidly with accruing data, treatment appear more efficacious. because there was no effective vaccine but also avoiding the long lead times and substantial A recent study used a statistical approach that because of concerns over increasing drug resist- costs associated with randomised trials. They claimed to overcome survivorship bias by includ- ance to amantadine and rimatidine.3 can also be conducted when it is considered ing treatment with antivirals as a time depend- unethical to randomise patients between treat- ent explanatory variable.6 This approach can Why randomised trials aren’t the be all and ment conditions, such as in pregnancy.4 Obser- address the survivorship bias associated with end all vational studies may also provide information early deaths, but the method does nothing to Randomised controlled trials are the gold stand- on the safety of interventions because they can recover information about patients missing from ard in research and have a central role in drug include more participants and follow them up the analysis. evaluation and health policy. Randomisation over a longer period. Statistical techniques are available to address ensures that participants differ only by the play The challenge for observational studies is that other types of confounding in observational stud- of chance and treatment allocation. This means treatment is not allocated by the play of chance ies, including multivariable statistical modelling that in a well designed randomised trial, the only (randomisation) so they are subject to substan- and propensity scores.4 Propensity scores, which explanations for a difference in outcome are the tial bias.4 Observational studies should adopt use patient characteristics to estimate the likeli- effect of the experimental treatment or chance, approaches to reduce bias and potential con- hood they received a certain treatment, can be and if chance is not a plausible explanation (for founders, but they can account only for known used to match groups or included in a multivari- 22 BMJ | 12 APRIL 2014 | VOLUME 348 ANALYSIS able model to adjust for or weight different pro- Identification of observational studies 7 pensities to receiving the treatments. Propensity Search methods scores are not as robust as randomisation, as they We designed a sensitive search to retrieve observational studies from electronic bibliographic incorporate only available risk information and databases. In order to retrieve non-randomised studies, we used no study design filter. No date often in a somewhat crude way. In the observa- limitation or language restriction was applied. tional studies of oseltamivir, propensity scores We identified 7523 items from the following databases on 25 February 2014 adjusting for patient risk may not account for MEDLINE via OVID (1946 to 25 February 2013) other factors such as patient entitlement to care, EMBASE via OVID (1947 to 24 Feb 2013) staffing levels, or high dependency care facilities, Cochrane Library via Wiley (Issue 2 of 12, 2014) including the Database of Reviews of Effects (DARE), Health Technology Assessment (HTA), and Economic Evaluations Database (EED) to scan the reference which can partly explain the outcomes measured. lists of relevant systematic reviews PUBMED via NLM ‘Related Articles’ search in PUBMED using the previous review of observational Taking stock of the evidence studies (Freemantle, 2009) as a seed paper2 Following the 2009 H1N1 pandemic the World The search strategy was devised on OVID MEDLINE and then adapted for the other databases. The Health Organization commissioned Hsu and col- search strategies are available in appendix 1 on bmj.com. Screening of the search was limited to studies leagues to review the observational evidence for published from January 2010 to update the searches conducted by Hsu et al.8 In addition we screened oseltamivir, including studies up to November the 51 studies included in Hsu et al to see whether they would indeed meet our more stringent criteria. 8 2010. They found some evidence that oseltami- Inclusion and exclusion criteria vir might reduce mortality in high risk popula- We included studies published in English that compared outcomes for patients prescribed oseltamivir tions (odds ratio=0.23, 95% confidence interval versus patients prescribed no drug. Patients of all ages with or without comorbidities were eligible for 0.13 to 0.43) but they included low quality stud- inclusion. Randomised trials were excluded, as were observational studies without a “no treatment” ies in which the investigators did not use best comparator population. We excluded studies that did not use multivariable models or other methods to address bias. This makes the finding appropriate methods to condition for observed confounders. Thus, studies that use propensity score matching or adjustment were included, but studies that adjusted for a single confounding factor (age), of these studies unreliable, a point recognised or undertook no adjustment, were excluded. by the authors, leaving it unclear whether the pooled results are reliable. Study assessment and review methods DK and LS screened the titles and abstracts of publications identified in the search for potential We performed a systematic search of pub- inclusion. MC, NF, and LS screened full text articles.
Recommended publications
  • Pat Croskerry MD Phd
    Thinking (and the factors that influence it) Pat Croskerry MD PhD Scottish Intensive Care Society St Andrews, January 2011 RECOGNIZED Intuition Pattern Pattern Initial Recognition Executive Dysrationalia Calibration Output information Processor override T override Repetition NOT Analytical RECOGNIZED reasoning Dual Process Model Medical Decision Making Intuitive Analytical Orthodox Medical Decision Making (Analytical) Rational Medical Decision-Making • Knowledge base • Differential diagnosis • Best evidence • Reviews, meta-analysis • Biostatistics • Publication bias, citation bias • Test selection and interpretation • Bayesian reasoning • Hypothetico-deductive reasoning .„Cognitive thought is the tip of an enormous iceberg. It is the rule of thumb among cognitive scientists that unconscious thought is 95% of all thought – .this 95% below the surface of conscious awareness shapes and structures all conscious thought‟ Lakoff and Johnson, 1999 Rational blind-spots • Framing • Context • Ambient conditions • Individual factors Individual Factors • Knowledge • Intellect • Personality • Critical thinking ability • Decision making style • Gender • Ageing • Circadian type • Affective state • Fatigue, sleep deprivation, sleep debt • Cognitive load tolerance • Susceptibility to group pressures • Deference to authority Intelligence • Measurement of intelligence? • IQ most widely used barometer of intellect and cognitive functioning • IQ is strongest single predictor of job performance and success • IQ tests highly correlated with each other • Population
    [Show full text]
  • A Task-Based Taxonomy of Cognitive Biases for Information Visualization
    A Task-based Taxonomy of Cognitive Biases for Information Visualization Evanthia Dimara, Steven Franconeri, Catherine Plaisant, Anastasia Bezerianos, and Pierre Dragicevic Three kinds of limitations The Computer The Display 2 Three kinds of limitations The Computer The Display The Human 3 Three kinds of limitations: humans • Human vision ️ has limitations • Human reasoning 易 has limitations The Human 4 ️Perceptual bias Magnitude estimation 5 ️Perceptual bias Magnitude estimation Color perception 6 易 Cognitive bias Behaviors when humans consistently behave irrationally Pohl’s criteria distilled: • Are predictable and consistent • People are unaware they’re doing them • Are not misunderstandings 7 Ambiguity effect, Anchoring or focalism, Anthropocentric thinking, Anthropomorphism or personification, Attentional bias, Attribute substitution, Automation bias, Availability heuristic, Availability cascade, Backfire effect, Bandwagon effect, Base rate fallacy or Base rate neglect, Belief bias, Ben Franklin effect, Berkson's paradox, Bias blind spot, Choice-supportive bias, Clustering illusion, Compassion fade, Confirmation bias, Congruence bias, Conjunction fallacy, Conservatism (belief revision), Continued influence effect, Contrast effect, Courtesy bias, Curse of knowledge, Declinism, Decoy effect, Default effect, Denomination effect, Disposition effect, Distinction bias, Dread aversion, Dunning–Kruger effect, Duration neglect, Empathy gap, End-of-history illusion, Endowment effect, Exaggerated expectation, Experimenter's or expectation bias,
    [Show full text]
  • Cognitive Biases in Software Engineering: a Systematic Mapping Study
    Cognitive Biases in Software Engineering: A Systematic Mapping Study Rahul Mohanani, Iflaah Salman, Burak Turhan, Member, IEEE, Pilar Rodriguez and Paul Ralph Abstract—One source of software project challenges and failures is the systematic errors introduced by human cognitive biases. Although extensively explored in cognitive psychology, investigations concerning cognitive biases have only recently gained popularity in software engineering research. This paper therefore systematically maps, aggregates and synthesizes the literature on cognitive biases in software engineering to generate a comprehensive body of knowledge, understand state of the art research and provide guidelines for future research and practise. Focusing on bias antecedents, effects and mitigation techniques, we identified 65 articles (published between 1990 and 2016), which investigate 37 cognitive biases. Despite strong and increasing interest, the results reveal a scarcity of research on mitigation techniques and poor theoretical foundations in understanding and interpreting cognitive biases. Although bias-related research has generated many new insights in the software engineering community, specific bias mitigation techniques are still needed for software professionals to overcome the deleterious effects of cognitive biases on their work. Index Terms—Antecedents of cognitive bias. cognitive bias. debiasing, effects of cognitive bias. software engineering, systematic mapping. 1 INTRODUCTION OGNITIVE biases are systematic deviations from op- knowledge. No analogous review of SE research exists. The timal reasoning [1], [2]. In other words, they are re- purpose of this study is therefore as follows: curring errors in thinking, or patterns of bad judgment Purpose: to review, summarize and synthesize the current observable in different people and contexts. A well-known state of software engineering research involving cognitive example is confirmation bias—the tendency to pay more at- biases.
    [Show full text]
  • When Do Employees Perceive Their Skills to Be Firm-Specific?
    r Academy of Management Journal 2016, Vol. 59, No. 3, 766–790. http://dx.doi.org/10.5465/amj.2014.0286 MICRO-FOUNDATIONS OF FIRM-SPECIFIC HUMAN CAPITAL: WHEN DO EMPLOYEES PERCEIVE THEIR SKILLS TO BE FIRM-SPECIFIC? JOSEPH RAFFIEE University of Southern California RUSSELL COFF University of Wisconsin-Madison Drawing on human capital theory, strategy scholars have emphasized firm-specific human capital as a source of sustained competitive advantage. In this study, we begin to unpack the micro-foundations of firm-specific human capital by theoretically and empirically exploring when employees perceive their skills to be firm-specific. We first develop theoretical arguments and hypotheses based on the extant strategy literature, which implicitly assumes information efficiency and unbiased perceptions of firm- specificity. We then relax these assumptions and develop alternative hypotheses rooted in the cognitive psychology literature, which highlights biases in human judg- ment. We test our hypotheses using two data sources from Korea and the United States. Surprisingly, our results support the hypotheses based on cognitive bias—a stark contrast to expectations embedded within the strategy literature. Specifically, we find organizational commitment and, to some extent, tenure are negatively related to employee perceptions of the firm-specificity. We also find that employer-provided on- the-job training is unrelated to perceived firm-specificity. These results suggest that firm-specific human capital, as perceived by employees, may drive behavior in ways unanticipated by existing theory—for example, with respect to investments in skills or turnover decisions. This, in turn, may challenge the assumed relationship between firm-specific human capital and sustained competitive advantage.
    [Show full text]
  • Survivorship Bias Mitigation in a Recidivism Prediction Tool
    EasyChair Preprint № 4903 Survivorship Bias Mitigation in a Recidivism Prediction Tool Ninande Vermeer, Alexander Boer and Radboud Winkels EasyChair preprints are intended for rapid dissemination of research results and are integrated with the rest of EasyChair. January 15, 2021 SURVIVORSHIP BIAS MITIGATION IN A RECIDIVISM PREDICTION TOOL Ninande Vermeer / Alexander Boer / Radboud Winkels FNWI, University of Amsterdam, Netherlands; [email protected] KPMG, Amstelveen, Netherlands; [email protected] PPLE College, University of Amsterdam, Netherlands; [email protected] Keywords: survivorship bias, AI Fairness 360, bias mitigation, recidivism Abstract: Survivorship bias is the fallacy of focusing on entities that survived a certain selection process and overlooking the entities that did not. This common form of bias can lead to wrong conclu- sions. AI Fairness 360 is an open-source toolkit that can detect and handle bias using several mitigation techniques. However, what if the bias in the dataset is not bias, but rather a justified unbalance? Bias mitigation while the ‘bias’ is justified is undesirable, since it can have a serious negative impact on the performance of a prediction tool based on machine learning. In order to make well-informed product design decisions, it would be appealing to be able to run simulations of bias mitigation in several situations to explore its impact. This paper de- scribes the first results in creating such a tool for a recidivism prediction tool. The main con- tribution is an indication of the challenges that come with the creation of such a simulation tool, specifically a realistic dataset. 1. Introduction The substitution of human decision making with Artificial Intelligence (AI) technologies increases the depend- ence on trustworthy datasets.
    [Show full text]
  • Why Too Many Political Science Findings Cannot
    www.ssoar.info Why Too Many Political Science Findings Cannot be Trusted and What We Can Do About it: A Review of Meta-scientific Research and a Call for Institutional Reform Wuttke, Alexander Preprint / Preprint Zeitschriftenartikel / journal article Empfohlene Zitierung / Suggested Citation: Wuttke, A. (2019). Why Too Many Political Science Findings Cannot be Trusted and What We Can Do About it: A Review of Meta-scientific Research and a Call for Institutional Reform. Politische Vierteljahresschrift, 60(1). https:// doi.org/10.1007/s11615-018-0131-7 Nutzungsbedingungen: Terms of use: Dieser Text wird unter einer CC BY-NC Lizenz (Namensnennung- This document is made available under a CC BY-NC Licence Nicht-kommerziell) zur Verfügung gestellt. Nähere Auskünfte zu (Attribution-NonCommercial). For more Information see: den CC-Lizenzen finden Sie hier: https://creativecommons.org/licenses/by-nc/4.0 https://creativecommons.org/licenses/by-nc/4.0/deed.de Diese Version ist zitierbar unter / This version is citable under: https://nbn-resolving.org/urn:nbn:de:0168-ssoar-59909-5 Wuttke (2019): Credibility of Political Science Findings Why Too Many Political Science Findings Cannot be Trusted and What We Can Do About it: A Review of Meta-scientific Research and a Call for Institutional Reform Alexander Wuttke, University of Mannheim 2019, Politische Vierteljahresschrift / German Political Science Quarterly 1, 60: 1-22, DOI: 10.1007/s11615-018-0131-7. This is an uncorrected pre-print. Please cite the original article. Witnessing the ongoing “credibility revolutions” in other disciplines, also political science should engage in meta-scientific introspection. Theoretically, this commentary describes why scientists in academia’s current incentive system work against their self-interest if they prioritize research credibility.
    [Show full text]
  • Understanding the Replication Crisis As a Base Rate Fallacy
    Understanding the Replication Crisis as a Base-Rate Fallacy PhilInBioMed, Université de Bordeaux 6 June 2018 [email protected] introduction — the replication crisis 52% of 1,576 scientists taking a survey conducted by the journal Nature agreed that there was a significant crisis of reproducibility Amgen successfully replicated only 6 out of 53 studies in oncology And then there is social psychology . introduction — the base rate fallacy screening for a disease, which affects 1 in every 1,000 individuals, with a 95% accurate test an individual S tests positive, no other risk factors; what is the probability that S has the disease? Harvard medical students, 1978 11 out of 60 got the correct answer introduction — the base rate fallacy introduction — the base rate fallacy base rate of disease = 1 in 1,000 = 0.1% (call this π) false positive rate = 5% (call this α) false positives among the 999 disease-free greatly outnumber the 1 true positive from the base rate fallacy to the replication crisis two types of error and accuracy type of error error rate accuracy type of accuracy Type-I (false +ve) α 1– α confidence level Type-II (false −ve) β 1– β power from the base rate fallacy to the replication crisis do not conflate False Positive Report Probability (FPRP) Pr (S does not have the disease, given that S tests positive for the disease) with False positive error rate (α) Pr (S tests positive for the disease, given that S does not have the disease) from the base rate fallacy to the replication crisis do not conflate Pr (the temperature will
    [Show full text]
  • Perioperative Management of ACE Inhibitor Therapy: Challenges of Clinical Decision Making Based on Surrogate Endpoints
    EDITORIAL Perioperative Management of ACE Inhibitor Therapy: Challenges of Clinical Decision Making Based on Surrogate Endpoints Duminda N. Wijeysundera MD, PhD, FRCPC1,2,3* 1Li Ka Shing Knowledge Institute, St. Michael’s Hospital, Toronto, Ontario, Canada; 2Department of Anesthesia and Pain Management, Toronto General Hospital, Toronto, Ontario, Canada; 3Department of Anesthesia, University of Toronto, Toronto, Ontario, Canada. enin-angiotensin inhibitors, which include angio- major surgery can have adverse effects. While ACE inhibitor tensin-converting enzyme (ACE) inhibitors and an- withdrawal has not shown adverse physiological effects in the giotensin II receptor blockers (ARBs), have demon- perioperative setting, it has led to rebound myocardial isch- strated benefits in the treatment of several common emia in patients with prior myocardial infarction.6 Rcardiovascular and renal conditions. For example, they are Given this controversy, there is variation across hospitals1 prescribed to individuals with hypertension, heart failure with and practice guidelines with respect to perioperative manage- reduced ejection fraction (HFrEF), prior myocardial infarction, ment of renin-angiotensin inhibitors. For example, the 2017 and chronic kidney disease with proteinuria. Perhaps unsur- Canadian Cardiovascular Society guidelines recommend that prisingly, many individuals presenting for surgery are already renin-angiotensin inhibitors be stopped temporarily 24 hours on long-term ACE inhibitor or ARB therapy. For example, such before major inpatient
    [Show full text]
  • The Existence of Publication Bias and Risk Factors for Its Occurrence
    The Existence of Publication Bias and Risk Factors for Its Occurrence Kay Dickersin, PhD Publication bias is the tendency on the parts of investigators, reviewers, and lication, the bias in terms ofthe informa¬ editors to submit or accept manuscripts for publication based on the direction or tion available to the scientific communi¬ strength of the study findings. Much of what has been learned about publication ty may be considerable. The bias that is when of re¬ bias comes from the social sciences, less from the field of medicine. In medicine, created publication study sults is on the direction or three studies have direct evidence for this bias. Prevention of based signifi¬ provided publica- cance of the is called tion bias is both from the scientific dissemina- findings publica¬ important perspective (complete tion bias. This term seems to have been tion of knowledge) and from the perspective of those who combine results from a used first in the published scientific lit¬ number of similar studies (meta-analysis). If treatment decisions are based on erature by Smith1 in 1980. the published literature, then the literature must include all available data that is Even if publication bias exists, is it of acceptable quality. Currently, obtaining information regarding all studies worth worrying about? In a scholarly undertaken in a given field is difficult, even impossible. Registration of clinical sense, it is certainly worth worrying trials, and perhaps other types of studies, is the direction in which the scientific about. If one believes that judgments should move. about medical treatment should be community all (JAMA.
    [Show full text]
  • Assessing SRI Fund Performance Research: Best Practices in Empirical Analysis
    Sustainable Development Sust. Dev. (2011) Published online in Wiley Online Library (wileyonlinelibrary.com) DOI: 10.1002/sd.509 Assessing SRI Fund Performance Research: Best Practices in Empirical Analysis Andrea Chegut,1* Hans Schenk2 and Bert Scholtens3 1 Department of Finance, Maastricht University, Maastricht, The Netherlands 2 Department of Economics, Utrecht University, Utrecht, The Netherlands 3 Department of Economics, Econometrics and Finance, University of Groningen, Groningen, The Netherlands ABSTRACT We review the socially responsible investment (SRI) mutual fund performance literature to provide best practices in SRI performance attribution analysis. Based on meta-ethnography and content analysis, five themes in this literature require specific attention: data quality, social responsibility verification, survivorship bias, benchmarking, and sensitivity and robustness checks. For each of these themes, we develop best practices. Specifically, for sound SRI fund performance analysis, it is important that research pays attention to divi- dend yields and fees, incorporates independent and third party social responsibility verifica- tion, corrects for survivorship bias and tests multiple benchmarks, as well as analyzing the impact of fund composition, management influences and SRI strategies through sensitivity and robustness analysis. These best practices aim to enhance the robustness of SRI finan- cial performance analysis. Copyright © 2011 John Wiley & Sons, Ltd and ERP Environment. Received 1 September 2009; revised 2 December 2009; accepted 4 January 2010 Keywords: mutual funds; socially responsible investing; performance evaluation; best practices Introduction N THIS PAPER, WE INVESTIGATE PERFORMANCE ATTRIBUTION ANALYSIS WITH RESPECT TO SOCIALLY RESPONSIBLE investment (SRI). This analysis is relevant in the decision making process of financial institutions in construct- ing and offering SRI portfolios.
    [Show full text]
  • A Meta-Meta-Analysis
    Journal of Intelligence Article Effect Sizes, Power, and Biases in Intelligence Research: A Meta-Meta-Analysis Michèle B. Nuijten 1,* , Marcel A. L. M. van Assen 1,2, Hilde E. M. Augusteijn 1, Elise A. V. Crompvoets 1 and Jelte M. Wicherts 1 1 Department of Methodology & Statistics, Tilburg School of Social and Behavioral Sciences, Tilburg University, Warandelaan 2, 5037 AB Tilburg, The Netherlands; [email protected] (M.A.L.M.v.A.); [email protected] (H.E.M.A.); [email protected] (E.A.V.C.); [email protected] (J.M.W.) 2 Section Sociology, Faculty of Social and Behavioral Sciences, Utrecht University, Heidelberglaan 1, 3584 CS Utrecht, The Netherlands * Correspondence: [email protected]; Tel.: +31-13-466-2053 Received: 7 May 2020; Accepted: 24 September 2020; Published: 2 October 2020 Abstract: In this meta-study, we analyzed 2442 effect sizes from 131 meta-analyses in intelligence research, published from 1984 to 2014, to estimate the average effect size, median power, and evidence for bias. We found that the average effect size in intelligence research was a Pearson’s correlation of 0.26, and the median sample size was 60. Furthermore, across primary studies, we found a median power of 11.9% to detect a small effect, 54.5% to detect a medium effect, and 93.9% to detect a large effect. We documented differences in average effect size and median estimated power between different types of intelligence studies (correlational studies, studies of group differences, experiments, toxicology, and behavior genetics).
    [Show full text]
  • Protecting Against Researcher Bias in Secondary Data Analysis
    Protecting against researcher bias in secondary data analysis: Challenges and potential solutions Jessie R. Baldwin1,2, PhD, Jean-Baptiste Pingault1,2, PhD, Tabea Schoeler,1 PhD, Hannah M. Sallis3,4,5, PhD & Marcus R. Munafò3,4,6, PhD 4,618 words; 2 tables; 1 figure 1 Department of Clinical, Educational and Health Psychology, Division of Psychology and Language Sciences, University College London, London, UK 2 Social, Genetic and Developmental Psychiatry Centre, Institute of Psychiatry, Psychology and Neuroscience, King’s College London, London, UK 3 MRC Integrative Epidemiology Unit at the University of Bristol, Bristol Medical School, University of Bristol, Bristol, UK 4 School of Psychological Science, University of Bristol, Bristol, UK 5 Centre for Academic Mental Health, Population Health Sciences, University of Bristol, Bristol, UK 6 NIHR Biomedical Research Centre, University Hospitals Bristol NHS Foundation Trust and University of Bristol, Bristol, UK Correspondence: Dr. Jessie R. Baldwin, Department of Clinical, Educational and Health Psychology, Division of Psychology and Language Sciences, University College London, London, WC1H 0AP, UK; [email protected] Funding: J.R.B is funded by a Wellcome Trust Sir Henry Wellcome fellowship (grant 215917/Z/19/Z). J.B.P is a supported by the Medical Research Foundation 2018 Emerging Leaders 1st Prize in Adolescent Mental Health (MRF-160-0002-ELP-PINGA). M.R.M and H.M.S work in a unit that receives funding from the University of Bristol and the UK Medical Research Council (MC_UU_00011/5, MC_UU_00011/7), and M.R.M is also supported by the National Institute for Health Research (NIHR) Biomedical Research Centre at the University Hospitals Bristol National Health Service Foundation Trust and the University of Bristol.
    [Show full text]