The Map of Organisation and People Involved in X

Total Page:16

File Type:pdf, Size:1020Kb

The Map of Organisation and People Involved in X Organizations, sites and people involved in x-risks prevention Global Impact Nano Size and level of AI risks General x-risks Nuclear Bio-risks influence warming risks risks FHI Future of humanity institute Bulletin of IPCC World Health NASA Foresight Known MIRI FLI Club of Rome Organization link (Former Singularity Future of life institute Oxford, link Still exist! Atomic International (WHO) institute very well institute) Elon Musk Nick Bostrom Were famous in 1970s when they pro- panel of climate and large Scientists change includes a division called the Global link E.Yudkowsky link list of researchers duced “Limits of growth” Link Alert and Response (GAR) which mon- Famous doomsday itors and responds to global epidemic amount of work link crisis. GAR helps member states with clock training and coordination of response to is done epidemics. link link OpenAI Oxford Martin CSER EA B612 Elon Musk Programme Cambridge center of existential risks The United States Effective altruism on the Impacts of Future wiki Martin Rees, link The Forum Agency for Interna- Foundation Open Philanthropy Technology, link for Climate tional Development project (USAID) 80’000 hours Nuclear Engineering has its Emerging Pandemic Threats Program which aims to prevent and threat Initia- Assessment contain naturally generated pandemics Got grant from OFP at their source.[129] Global Priorities X-risks GCRI Foundational Skoll Global CISAC tive link Large and Global catastrophic risks Project institute institute, Research Threats Fund “The Center for Internation- interest- Phil Torrres Seth Baum al Security and Cooperation is Created Global catastrophic risk re- Institute “To safeguard humanity from Stanford University’s hub for re- port-2016 Interesting articles by its Interesting newsletter, many “Currently, our research focus- global threats” searchers tackling some of the Arctic news The Lawrence ing Collaborate with UK government main author, focus on exis- articles in scientific journals es on reducing risks of dysto- Climate, water security, pandem- world’s most pressing securi- Stimson Livermore National Holocene Dr. Toby Ord is member tential terrorism and religion pian futures in the context of ics, nuclear proliferation, link Sam Carana link ty and international cooperation Irreversible global warming “Global Priorities Project has link emerging technologies”. problems” Laboratory Impact Working research Center because of methane hydrates has a division called the Global Secu- merged into Centre for Effective Al- Interesting work on AI safety Nuclear, cybersecurity, bio, anti- “The Stimson Center is a eruption rity Principal Directorate which re- Group has been cre- truism (see here). We are continu- searches on behalf of the government link terrorism, link nonpartisan policy research link Estimate risks of recent ing to do research on questions re- issues such as bio-security, counter-ter- ated, but not center working to solve the rorism, etc. Link impacts lated to existential risk, though we world’s greatest threats to link much people are not currently planning to write security and prosperity”. knows them new reports on the topic like the non-prolifiration report mentioned above.” X-risks Leverage Convergence Lifeboat Saving link link Net Research Justin Shovelain Foundation Humanity Alexei Turchin Collective think tank concentrat- Very large scientific boards which Creating full database on the site is almost empty now ed on mathematical modeling of don’t actually do anything, but from Homo Sapiens Small one person organisation flutrackers.com x-risks and prevention plan link x-risks some useful discussion is going Forum about risks of flu pandemic without any actual work link in its mail list link link site Ploughshares UC Berkeley — Center mail list - good one Fund for Human- “Supports the smartest minds and most effective or- Compatible AI ganizations to reduce nucle- link ar stockpiles, prevent new Leverhulme Centre for Global nuclear states, and increase iGEM — Synthet- the Future of Intelli- global security” ic Biology Safety challenges Link gence Laszlo Szombatfalvy and Security The Global Challenges Foundation works to AI Safety raise awareness of the Global Catastroph- Aarhus ic Risks. Primarily focused on climate change, Denmark other environmental degradation and politi- cally motivated violence as well as how these Diffusing Nuclear link threats are linked to poverty and rapid popu- Threat Blue Ribbon lation growth, link link Study Panel on Biodefense Jaan Tallinn Bill Gates Investors Elon Musk Sam Altman Bill Joy Peter Thiel Want AI safety through Open Y combinator, Wrote famous arti- and important investor in Has its own foun- DEFCON warning AI and human on Mars as a Confounded cle but now seems to Invested figures x-related dation and vision system backup plan Open AI lost interest in MIRI projects, wiki of global risks Public Zoltvan Stephen Istavn figures Hawking Warned about risks of Presidential candidate aliens and AI from transhumanist party Wrote about x-risks Writers Vernor Vinge Greg Egan David Brin John Barnes writer writer, “Mother of writer, created “Permutation Lewis Dartnell Singularity idea city” “Existence” storms” author of “Knowledge” A. Sandberg Adrian Toby Ord Robin Katja Grace Willard Wells R. Freitas Dennis Scientists Kent Hanson Meadows Participated in site Fermi paradox Author of “Apoc- Nanotech risks and re- Resource depletion risks FHI and co-au- LHC risks “existential Blog and DA alypses when” searchers thored papers hope” Societal collapse blog and prevention risks AI impacts plan David Alexander Nick Milan Ćirković Bruce Tonn Large group of people working Denkenberger Kononov Beackstead agricultural risks Stevenson probe, Editor and writer on AI safety, Coined term FHI, work on a problem Anthropic shadow link including, but not limited to: ”indestructibility of of refuges against x-risks Fermi paradox Steve Omohundro civilization” Site Luke Muehlhauser Stuart Armstrong Roman Yampolskiy Nate Soares R.Carrigan Max Tegmark Norwegian Vladimir Nesov Aaron Dar Bill Napier Wrote articles together transhumanists Kaj Sotala with Bostrom Adrian Rohreim Benja Fallenstein Risks of SETI Risks of Risks of dark Riva Melissa Tez supernovas comets Jason Gaverick Matheny, wiki Andrew Critch, blog Paul Christiano Carl Shulman Anna Salamon Open Less- Facebook X-risks Longecity Wiki- EA forum Intelligent Discussion in • Existential risks (Adam Ford) agents forum places for wrong • Global Catastrophic Risks Re- on subforum resources link Technical discussion comments search and Discussion (Evan IEET discussion Gaensbauer) Reddit on AI safety • Global catastrophic risks link LW-wiki link Futureoflife • Stop existential risks Hplus pedia Existentialrisk Control problem.
Recommended publications
  • TEDX – What's Happening with Artificial Intelligence
    What’s Happening With Artificial Intelligence? Steve Omohundro, Ph.D. PossibilityResearch.com SteveOmohundro.com SelfAwareSystems.com http://googleresearch.blogspot.com/2015/06/inceptionism-going-deeper-into-neural.html Multi-Billion Dollar Investments • 2013 Facebook – AI lab • 2013 Ebay – AI lab • 2013 Allen Institute for AI • 2014 IBM - $1 billion in Watson • 2014 Google - $500 million, DeepMind • 2014 Vicarious - $70 million • 2014 Microsoft – Project Adam, Cortana • 2014 Baidu – Silicon Valley • 2015 Fanuc – Machine Learning for Robotics • 2015 Toyota – $1 billion, Silicon Valley • 2016 OpenAI – $1 billion, Silicon Valley http://www.mckinsey.com/insights/business_technology/disruptive_technologies McKinsey: AI and Robotics to 2025 $50 Trillion! US GDP is $18 Trillion http://cdn-media-1.lifehack.org/wp-content/files/2014/07/Cash.jpg 86 Billion Neurons https://upload.wikimedia.org/wikipedia/commons/e/ef/Human_brain_01.jpg http://www.ncbi.nlm.nih.gov/pmc/articles/PMC2776484/ The Connectome http://discovermagazine.com/~/media/Images/Issues/2013/Jan-Feb/connectome.jpg 1957 Rosenblatt’s “Perceptron” http://www.rutherfordjournal.org/article040101.html http://bio3520.nicerweb.com/Locked/chap/ch03/3_11-neuron.jpg https://upload.wikimedia.org/wikipedia/commons/3/31/Perceptron.svg “The embryo of an electronic computer that [the Navy] expects will be able to walk, talk, see, write, reproduce itself and be conscious of its existence.” https://en.wikipedia.org/wiki/Perceptron 1986 Backpropagation http://www.ifp.illinois.edu/~yuhuang/samsung/ANN.png
    [Show full text]
  • Artificial Intelligence: Distinguishing Between Types & Definitions
    19 NEV. L.J. 1015, MARTINEZ 5/28/2019 10:48 AM ARTIFICIAL INTELLIGENCE: DISTINGUISHING BETWEEN TYPES & DEFINITIONS Rex Martinez* “We should make every effort to understand the new technology. We should take into account the possibility that developing technology may have im- portant societal implications that will become apparent only with time. We should not jump to the conclusion that new technology is fundamentally the same as some older thing with which we are familiar. And we should not hasti- ly dismiss the judgment of legislators, who may be in a better position than we are to assess the implications of new technology.”–Supreme Court Justice Samuel Alito1 TABLE OF CONTENTS INTRODUCTION ............................................................................................. 1016 I. WHY THIS MATTERS ......................................................................... 1018 II. WHAT IS ARTIFICIAL INTELLIGENCE? ............................................... 1023 A. The Development of Artificial Intelligence ............................... 1023 B. Computer Science Approaches to Artificial Intelligence .......... 1025 C. Autonomy .................................................................................. 1026 D. Strong AI & Weak AI ................................................................ 1027 III. CURRENT STATE OF AI DEFINITIONS ................................................ 1029 A. Black’s Law Dictionary ............................................................ 1029 B. Nevada .....................................................................................
    [Show full text]
  • Wise Leadership & AI 3
    Wise Leadership and AI Leadership Chapter 3 | Behind the Scenes of the Machines What’s Ahead For Artificial General Intelligence? By Dr. Peter VERHEZEN With the AMROP EDITORIAL BOARD Putting the G in AI | 8 points True, generalized intelligence will be achieved when computers can do or learn anything that a human can. At the highest level, this will mean that computers aren’t just able to process the ‘what,’ but understand the ‘why’ behind data — context, and cause and effect relationships. Even someday chieving consciousness. All of this will demand ethical and emotional intelligence. 1 We underestimate ourselves The human brain is amazingly general compared to any digital device yet developed. It processes bottom-up and top-down information, whereas AI (still) only works bottom-up, based on what it ‘sees’, working on specific, narrowly defined tasks. So, unlike humans, AI is not yet situationally aware, nuanced, or multi-dimensional. 2 When can we expect AGI? Great minds do not think alike Some eminent thinkers (and tech entrepreneurs) see true AGI as only a decade or two away. Others see it as science fiction — AI will more likely serve to amplify human intelligence, just as mechanical machines have amplified physical strength. 3 AGI means moving from homo sapiens to homo deus Reaching AGI has been described by the futurist Ray Kurzweil as ‘singularity’. At this point, humans should progress to the ‘trans-human’ stage: cyber-humans (electronically enhanced) or neuro-augmented (bio-genetically enhanced). 4 The real risk with AGI is not malice, but unguided brilliance A super-intelligent machine will be fantastically good at meeting its goals.
    [Show full text]
  • An Open Letter to the United Nations Convention on Certain Conventional Weapons
    An Open Letter to the United Nations Convention on Certain Conventional Weapons As companies building the technologies in Artificial Intelligence and Robotics that may be repurposed to develop autonomous weapons, we feel especially responsible in raising this alarm. We warmly welcome the decision of the UN’s Conference of the Convention on Certain Conventional Weapons (CCW) to establish a Group of Governmental Experts (GGE) on Lethal Autonomous Weapon Systems. Many of our researchers and engineers are eager to offer technical advice to your deliberations. We commend the appointment of Ambassador Amandeep Singh Gill of India as chair of the GGE. We entreat the High Contracting Parties participating in the GGE to work hard at finding means to prevent an arms race in these weapons, to protect civilians from their misuse, and to avoid the destabilizing effects of these technologies. We regret that the GGE’s first meeting, which was due to start today, has been cancelled due to a small number of states failing to pay their financial contributions to the UN. We urge the High Contracting Parties therefore to double their efforts at the first meeting of the GGE now planned for November. Lethal autonomous weapons threaten to become the third revolution in warfare. Once developed, they will permit armed conflict to be fought at a scale greater than ever, and at timescales faster than humans can comprehend. These can be weapons of terror, weapons that despots and terrorists use against innocent populations, and weapons hacked to behave in undesirable ways. We do not have long to act. Once this Pandora’s box is opened, it will be hard to close.
    [Show full text]
  • Between Ape and Artilect Createspace V2
    Between Ape and Artilect Conversations with Pioneers of Artificial General Intelligence and Other Transformative Technologies Interviews Conducted and Edited by Ben Goertzel This work is offered under the following license terms: Creative Commons: Attribution-NonCommercial-NoDerivs 3.0 Unported (CC-BY-NC-ND-3.0) See http://creativecommons.org/licenses/by-nc-nd/3.0/ for details Copyright © 2013 Ben Goertzel All rights reserved. ISBN: ISBN-13: “Man is a rope stretched between the animal and the Superman – a rope over an abyss.” -- Friedrich Nietzsche, Thus Spake Zarathustra Table&of&Contents& Introduction ........................................................................................................ 7! Itamar Arel: AGI via Deep Learning ................................................................. 11! Pei Wang: What Do You Mean by “AI”? .......................................................... 23! Joscha Bach: Understanding the Mind ........................................................... 39! Hugo DeGaris: Will There be Cyborgs? .......................................................... 51! DeGaris Interviews Goertzel: Seeking the Sputnik of AGI .............................. 61! Linas Vepstas: AGI, Open Source and Our Economic Future ........................ 89! Joel Pitt: The Benefits of Open Source for AGI ............................................ 101! Randal Koene: Substrate-Independent Minds .............................................. 107! João Pedro de Magalhães: Ending Aging ....................................................
    [Show full text]
  • Letters to the Editor
    Articles Letters to the Editor Research Priorities for is a product of human intelligence; we puter scientists, innovators, entrepre - cannot predict what we might achieve neurs, statisti cians, journalists, engi - Robust and Beneficial when this intelligence is magnified by neers, authors, professors, teachers, stu - Artificial Intelligence: the tools AI may provide, but the eradi - dents, CEOs, economists, developers, An Open Letter cation of disease and poverty are not philosophers, artists, futurists, physi - unfathomable. Because of the great cists, filmmakers, health-care profes - rtificial intelligence (AI) research potential of AI, it is important to sionals, research analysts, and members Ahas explored a variety of problems research how to reap its benefits while of many other fields. The earliest signa - and approaches since its inception, but avoiding potential pitfalls. tories follow, reproduced in order and as for the last 20 years or so has been The progress in AI research makes it they signed. For the complete list, see focused on the problems surrounding timely to focus research not only on tinyurl.com/ailetter. - ed. the construction of intelligent agents making AI more capable, but also on Stuart Russell, Berkeley, Professor of Com - — systems that perceive and act in maximizing the societal benefit of AI. puter Science, director of the Center for some environment. In this context, Such considerations motivated the “intelligence” is related to statistical Intelligent Systems, and coauthor of the AAAI 2008–09 Presidential Panel on standard textbook Artificial Intelligence: a and economic notions of rationality — Long-Term AI Futures and other proj - Modern Approach colloquially, the ability to make good ects on AI impacts, and constitute a sig - Tom Dietterich, Oregon State, President of decisions, plans, or inferences.
    [Show full text]
  • On the Danger of Artificial Intelligence
    On The Danger of Artificial Intelligence Saba Samiei A thesis submitted to Auckland University of Technology in fulfilment of the requirements for the degree of Master of Computing and Information Sciences (MCIS) July 2019 School of Engineering, Computer and Mathematical Sciences 1 | P a g e Abstract In 2017, the world economic forum announced that AI would increase the global economy by USD 16 trillion by 2030 (World Economic Forum, 2017). Yet, at the same time, some of the world’s most influential leaders warned us about the danger of AI. Is AI good or bad? Of utmost importance, is AI an existential threat to humanity? This thesis examines the latter question by breaking it down into three sub-questions, is the danger real?, is the defence adequate?, and how a doomsday scenario could happen?, and critically reviewing the literature in search for an answer. If true, and sadly it is, I conclude that AI is an existential threat to humanity. The arguments are as follows. The current rapid developments of robots, the success of machine learning, and the emergence of highly profitable AI companies will guarantee the rise of the machines among us. Sadly, among them are machines that are destructive, and the danger becomes real. A review of current ideas preventing such a doomsday event is, however, shown to be inadequate and a futuristic look at how doomsday could emerge is, unfortunately, promising! Keywords: AI, artificial intelligence, ethics, the danger of AI. 2 | P a g e Acknowledgements No work of art, science, anything in between or beyond is possible without the help of those currently around us and those who have previously laid the foundation of success for us.
    [Show full text]
  • AIDA Hearing on AI and Competitiveness of 23 March 2021
    SPECIAL COMMITTEE ON ARTIFICIAL INTELLIGENCE IN A DIGITAL AGE (AIDA) HEARING ON AI AND COMPETITIVENESS Panel I: AI Governance Kristi Talving, Deputy Secretary General for Business Environment, Ministry of Economic Affairs and Communications, Estonia Khalil Rouhana, Deputy Director General DG-CONNECT (CNECT), European Commission Kay Firth-Butterfield, Head of Artificial Intelligence and Machine Learnings; Member of the Executive Committee, World Economic Forum Dr. Sebastian Wieczorek, Vice President – Artificial Intelligence Technology, SAP SE, external expert (until October 2020) in the study commission on AI in the Bundestag * * * Panel II: the perspective of Business and the Industry Prof. Volker Markl, Chair of Research Group at TU Berlin, Database Systems and Information Management, Director of the Intelligent Analytics for Massive Data Research Group at DFKI and Director of the Berlin Big Data Center and Secretary of the VLDB Endowment Moojan Asghari, Cofounder/Initiator of Women in AI, Founder/CEO of Thousand Eyes On Me Marina Geymonat, Expert for AI strategy @ Ministry for Economic Development, Italy. Head, Artificial intelligence Platform @TIM, Telecom Italia Group Jaan Tallinn, founding engineer of Skype and Kazaa as well as a cofounder of the Cambridge Centre for the Study of Existential Risk and Future of Life Institute 2 23-03-2021 BRUSSELS TUESDAY 23 MARCH 2021 1-002-0000 IN THE CHAIR: DRAGOŞ TUDORACHE Chair of the Special Committee on Artificial Intelligence in a Digital Age (The meeting opened at 9.06) Opening remarks 1-003-0000 Chair. – Good morning dear colleagues. I hope you are all connected and you can hear and see us in the room. Welcome to this new hearing of our committee.
    [Show full text]
  • Ssoar-2020-Selle-Der Effektive Altruismus Als Neue.Pdf
    www.ssoar.info Der effektive Altruismus als neue Größe auf dem deutschen Spendenmarkt: Analyse von Spendermotivation und Leistungsmerkmalen von Nichtregierungsorganisationen (NRO) auf das Spenderverhalten; eine Handlungsempfehlung für klassische NRO Selle, Julia Veröffentlichungsversion / Published Version Arbeitspapier / working paper Empfohlene Zitierung / Suggested Citation: Selle, J. (2020). Der effektive Altruismus als neue Größe auf dem deutschen Spendenmarkt: Analyse von Spendermotivation und Leistungsmerkmalen von Nichtregierungsorganisationen (NRO) auf das Spenderverhalten; eine Handlungsempfehlung für klassische NRO. (Opuscula, 137). Berlin: Maecenata Institut für Philanthropie und Zivilgesellschaft. https://nbn-resolving.org/urn:nbn:de:0168-ssoar-67950-4 Nutzungsbedingungen: Terms of use: Dieser Text wird unter einer CC BY-NC-ND Lizenz This document is made available under a CC BY-NC-ND Licence (Namensnennung-Nicht-kommerziell-Keine Bearbeitung) zur (Attribution-Non Comercial-NoDerivatives). For more Information Verfügung gestellt. Nähere Auskünfte zu den CC-Lizenzen finden see: Sie hier: https://creativecommons.org/licenses/by-nc-nd/3.0 https://creativecommons.org/licenses/by-nc-nd/3.0/deed.de MAECENATA Julia Selle Der effektive Altruismus als neue Größe auf dem deutschen Spendenmarkt Analyse von Spendermotivation und Leistungsmerkmalen von Nichtregierungsorganisationen (NRO) auf das Spenderverhalten. Eine Handlungsempfehlung für klassische NRO. Opusculum Nr.137 Juni 2020 Die Autorin Julia Selle studierte an den Universität
    [Show full text]
  • Comments to Michael Jackson's Keynote on Determining Energy
    Comments to Michael Jackson’s Keynote on Determining Energy Futures using Artificial Intelligence Prof Sirkka Heinonen Finland Futures Research Centre (FFRC) University of Turku ENERGIZING FUTURES 13–14 June 2018 Tampere, Finland AI and Energy • Concrete tools: How Shaping Tomorrow answers the question How can AI help? • Goal of foresight crystal clear: Making better decisions today Huge Challenge – The Challenge The world will need to cut energy-related carbon dioxide emissions by 60 percent by 2050 -even as the population grows by more than two billion people Bold Solution on the Horizon The Renewable Energy Transition Companies as pioneers on energy themes, demand, supply, consumption • Google will reach 100% RE for its global operations this year • GE using AI to boost different forms of energy production and use tech-driven data reports to anticipate performance and maintenance needs around the world BUT …also the role of governments, cities and citizens …NGOs, media… new actors should be emphasised AI + Energy + Peer-to-Peer Society • The transformation of the energy system aligns with the principles of a Peer-to-Peer Society. • Peer-to-peer practices are based on the active participation and self-organisation of citizens. Citizens share knowledge, skills, co-create, and form new peer groups. • Citizens will use their capabilities also to develop energy-related products and services Rethinking Concepts Buildings as Power stations – global (economic) opportunity to construct buildings that generate, store and release solar energy
    [Show full text]
  • 2016 Annual Report
    Technology is giving life ...or to self-destruct. the potential to flourish Let’s make a difference! like never before... Annual Report Making a Difference 2016 1 2016 By the Numbers Nuclear Participants Conference 200 We launched our nuclear divestment campaign at Speakers MIT’S Nuclear Conference. 18 and panelists Billion dollars committed $1 to nuclear divestment Conferences attended AI Safety 17 by FLI members Events This is an increase of 36 AI safety events from the Invitations to speak at events previous year. 15 Events co-organized to 4 highlight AI safety ubshe, eerreee ers Grant Recipients 43 ur rnt reents re rey hr t or to ee Worshos rntees orne 87 nor rtte n bene. rtes rtten by or bout rntee reserh 16 or mor ubtons 2 2016 By the Numbers, Cont. on eoe the 1 rete eos Website We rehe oer mon on sts eoe onne bout .5 to s estent rs n hoe. on sts to .5 rtes n ebes rtes220 ubshe ntony tht reerene Months of News s mentone n mor nes outets eery month o the yer. nternton100 rtes tht reerene rtes11 rtten or the ste Tot1 number o ontons Donors Wee hee so muh thns to the enerosty o our onors. on$3.3 onte to roets n onors ho e t est8 , 3 Message from the President It’s been a great honor for me to get to work with such a talented and idealistic team at our institute to ensure that tomorrow’s most powerful technologies have a positive impact on humanity.
    [Show full text]
  • The Future of AI: Opportunities and Challenges
    The Future of AI: Opportunities and Challenges Puerto Rico, January 2-5, 2015 ! Ajay Agrawal is the Peter Munk Professor of Entrepreneurship at the University of Toronto's Rotman School of Management, Research Associate at the National Bureau of Economic Research in Cambridge, MA, Founder of the Creative Destruction Lab, and Co-founder of The Next 36. His research is focused on the economics of science and innovation. He serves on the editorial boards of Management Science, the Journal of Urban Economics, and The Strategic Management Journal. & Anthony Aguirre has worked on a wide variety of topics in theoretical cosmology, ranging from intergalactic dust to galaxy formation to gravity physics to the large-scale structure of inflationary universes and the arrow of time. He also has strong interest in science outreach, and has appeared in numerous science documentaries. He is a co-founder of the Foundational Questions Institute and the Future of Life Institute. & Geoff Anders is the founder of Leverage Research, a research institute that studies psychology, cognitive enhancement, scientific methodology, and the impact of technology on society. He is also a member of the Effective Altruism movement, a movement dedicated to improving the world in the most effective ways. Like many of the members of the Effective Altruism movement, Geoff is deeply interested in the potential impact of new technologies, especially artificial intelligence. & Blaise Agüera y Arcas works on machine learning at Google. Previously a Distinguished Engineer at Microsoft, he has worked on augmented reality, mapping, wearable computing and natural user interfaces. He was the co-creator of Photosynth, software that assembles photos into 3D environments.
    [Show full text]