Global Solutions vs. Local Solutions for the AI Safety Problem
Abstract
:1. Introduction
2. AI Safety Levels
- “Last man”: At least one human being survives creation of strong AI, for example, possibly as an upload, or in a “zoo”.
- “AI survivors”: A group of people survive and continue to exist after the AI creation, and may be able to rebuild human civilization to some extent. This may happen if the AI halts [13] or leaves Earth.
- “No AI”: Any outcome where global catastrophe connected with AI has not occurred because there is no AI to provoke the catastrophe. This is a world in which a comprehensive ban on AI is enforced, or AI technologies otherwise never progress to AGI or superintelligence.
- “Better now”: Human civilization is preserved after AI creation in almost the same form in which it exists now, and benefits from AI in many ways, including through the curing of diseases, slowing aging, preventing crime, increasing material goods, achieving interstellar travel, etc. Outcomes in this category likely involve a type of “AI Nanny” [21].
- “Infinite good”: Superintelligent AI which maximizes human values (Benevolent AI) will reach the maximum possible positive utility for humans, but contemporary humans cannot now describe this utility as it is beyond our ability to imagine, as presented by Yudkowsky [22].
3. “No AI” Solutions
- International ban
- Legal relinquishment
- Technical relinquishment or AI appears to be not technically possible
- Destruction of capability to produce AI anywhere in the world
- -
- War
- -
- Luddism
- -
- Staging small catastrophe
- Slowdown of AI creation
- -
- Economical
- -
- Technology slowdown
- -
- Overregulation
- -
- Brain drain from the field
- -
- Defamation of idea of AI, AI winter
3.1. Legal Solutions, Including Bans
3.2. Restriction Solutions
- -
- supercomputers
- -
- programmers
- -
- knowledge about AI creation
- -
- semiconductor fabrication plants (“fabs”)
- -
- internet access
- -
- electricity
3.3. Destructive Solutions
3.4. Delay of AI Creation
- Public fears of AI.
- The next AI winter, lack of interest in its development (there have already been two after hype in the 1960s and 1980s).
- Extensive regulation of the field.
- Intentional disruption of the research field via fake news, defamation, white noise, and other instruments of informational warfare.
- Public ridicule of the field after some failure.
- Change of focus of public attention by substitution of terms. This happened with “nanotechnology”, which originally meant a powerful manufacturing technology, but now means making anything small. Such a shift may happen with the term “AI,” where the meaning has shifted recently from human-like systems to narrow machine learning algorithms. There are several fields that have had slow development for decades because of marginalization, such as cryonics, but it looks as though the time of marginalization of AI has passed.
- Lastly, depending on the technical challenges, advanced AI, including AGI and superintelligence, may not be technically possible in the near future, although there is no reason at this point to assume it is not. However, if these challenges appear, AI could be indefinitely delayed.
4. ”One AI” Solutions
- First AI is used to take over the world
- ○
- First AI is used as a military instrument
- ○
- First AI gains global power via peaceful means
- -
- Commercial success
- -
- Superhuman negotiating abilities
- ○
- Strategic advantage achieved by narrow AIs produces global unification, before the rise of superintelligent AI, by leveraging preexisting advantage of a nuclear power and increasing first-strike capability
- ○
- First AI is created by a superpower and provides it a decisive strategic advantage
- ○
- First AI is reactive, and while it does not prevent the creation of other AI, it limits their potential danger
- ○
- First AI is a genius at negotiation and solves all conflicts between other agents
- First AI appears as a result of collective efforts
- ○
- AI police: global surveillance system to prevent creation of dangerous AI
- ○
- “AI CERN”: international collaboration creates an AI Nanny
- ○
- Main players collaborate with each other
- ○
- AIs are effective in cooperation and merge with each other
- Non-agential AI-medium (AI as widely distributed technology, without agency)
- ○
- Comprehensive AI Services
- ○
- Distributed AI based on blockchain (SingularityNET)
- ○
- AI as technology everywhere (openness)
- ○
- Augmented humans as AI neurons (Neuralink)
- ○
- Superintelligence as a distributed optimization process by rivalry between AI agents (market)
- Helping others to create safe first AI
- ○
- AI safety theory is distributed among main players and used by every AI creator
- ○
- AI safety instruments are sold as a service
- ○
- Promotion of AI safety
- Slowing creation of other AIs
- ○
- Concentrate best minds on other projects and remove them from AI research
- ○
- Take low-hanging research fruit
- Factors affecting the arms race for AI include funding, openness, number of teams, prizes, and public attitudes
4.1. First AI Seizes World Power
4.1.1. Concentrate the Best AI Researchers to Create a Powerful and Safe AI First
4.1.2. Using the Decisive Advantage of Non-Self-Improving AI to Create an AI Nanny
4.1.3. Risks of Creating Hard-takeoff AI as a Global Solution
4.2. One Global AI Created by Collective Efforts
4.2.1. AI Nanny Requires a World Government for Its Creation
- -
- General intelligence somewhat above the human level,
- -
- Interconnection with powerful worldwide surveillance systems,
- -
- Control of a massive contingent of robots, and
- -
- A cognitive architecture featuring an explicit set of goals.
4.2.2. Levels of Implementation of the AI Nanny Concept
4.2.3. Global Transition into AI: Non-Agential AI-Medium Everywhere, Accelerating Smoothly without Tipping Points
4.3. Help Others to Create Safe AI
4.3.1. Promoting Ideas of AI Safety in General and the Best AI Safety Solution to All Players
- -
- Funding of AI safety research.
- -
- Promotion of the idea of AI safety.
- -
- Protesting military AI.
- -
- Friendly AI training for AI researchers.
- -
- Providing publicly available safety recommendations.
- -
- Increasing the “sanity waterline” and rationality in the general population and among AI researchers and policymakers.
- -
- Lowering global levels of confrontation and enmity.
- -
- Forming political parties for the prevention of existential risks and control of AI risks, or lobbying current political parties to adopt these positions. However, even if such parties were to win in larger countries and were able to change policy, there would still be countries that could use any technology “freeze” in larger countries to their advantage.
- -
- Depleting the pool of minds for direct—not necessarily safe—AI research, thereby slowing it down
- -
- Increasing the quantity and quality of thought working on AI safety theory
- -
- Establishing relationships between the best AI teams, as some of the people who will have worked on AI safety may have come from such teams, may eventually join them, or may otherwise have friends there, and
- -
- Promoting the idea that unlimited self-improvement is dangerous and unstable for all players, including AIs.
4.3.2. Selling AI Safety Theory as an Effective Tool to Align Arbitrary AI
4.4. Local Action to Affect Other AIs Globally
4.4.1. Slowing the Appearance of Other AIs
4.4.2. Ways to Affect a Race to Create the First AI
- -
- Changing the number of participants.
- -
- Increasing or decreasing information exchange and level of openness.
- -
- Reducing the level of enmity between organizations and countries, and preventing conventional arms races and military buildups.
- -
- Increasing the level of cooperation, coordination, and acceptance of the idea of AI safety among AI researchers.
- -
- Changing the total amount of funding available.
- -
- Promoting intrinsic motivations for safety. Seth Baum discussed the weakness of monetary incentives for beneficial AI designs, and cautions: “One recurrent finding is that monetary incentives can reduce intrinsic motivation” [67]; when the money is gone, people lose motivation. Baum also noted that the mere fact that a law existed promoted obedience in some situations and that social encouragement can increase intrinsic motivation.
- -
- Changing social attitudes toward the problem and increasing awareness of the idea of AI safety.
- -
- Trying to affect the speed of the AI race, either slowing it down or accelerating it in just one place by concentrating research. It is interesting to note that acceleration could be done locally, but slowing it would require global cooperation, and so is less probable.
- -
- Affecting the idea of the AI race as it is understood by the participants [67]: if everybody thinks that the winner takes everything, the race is more dangerous. A similar framing solution has been suggested in the field of bioweapons, that is, to stop claiming bioweapon creation is easy, as it might become attractive to potential bioterrorists. In fact, bioweapons are not as easy to develop and deploy as is shown in movies, and would probably kill the terrorists first [68].
- -
- Affecting the public image of AI researchers who are currently presented as not wanting beneficial AI design [67].
- -
- Refraining from suggestions of draconian surveillance as they “inadvertently frame efforts to promote beneficial AI as being the problem, not the solution” [67].
- -
- Stigmatization of building recursive self-improving AI by framing them as morally unacceptable, as has been done with landmines. The stigma impelled even countries that did not sign the treaty that prohibits landmines to reduce production [67].
- -
- Deliberate association with crackpottery: an example is UFO (Unidentified Flying Objects) research: anyone who mentions the word “UFO” will no longer be accepted in the scientific community as a credible scientist. This partially worked against AI during past AI winters, when scientists tried not to mention the words “artificial intelligence.” Society could come to associate “self-improving AI” with craziness, which would be not difficult if we pick some of the most outstanding ideas from associated internet forums, e.g., Roko’s Basilisk [69]. Such an association may reduce funding for such research. However, AI could start to self-improve even if it was not designed to do so; thus, such association would probably be damaging to AI safety efforts. Recent successes in meta-learning in neural nets by DeepMind show that the idea of self-improving AI is becoming mainstream [70].
- -
- Affecting the speed of takeoff after one AI starts to win. If the speed of self-improvement of one AI diminishes, other AIs may catch up with it.
4.4.3. Participating in Acausal Deals with Future AI
5. “Many AI” Solutions
5.1. Overview of the “Net Solutions” of AI Safety
5.1.1. How a Net of AIs May Provide Global Safety
- The combined intelligence of many AIs (the net of AIs) is much higher than the one of any rogue AI, so the net is able to create effective protection. An AI-net could form something similar to AI “police,” which prevent any single AI from unlimited growth. This is analogous to the way the human body provides a multilevel defense against unlimited growth of a single cancerous cell in the form of an immune system. The approach is somewhat similar to the AI Nanny approach [21], but an AI Nanny is a single AI entity. An AI-net consists of many AIs, which use ubiquitous transparency [73] to control and balance [74] each other.
- Value diversity among many AI-sovereigns [2,75] guarantees that different positive values will not be lost. Different members of the net have different terminal values, thus ensuring diversity of values, as long as the values do not destructively interfere. If the values do destructively interfere, then solutions must be found for these conflicts.
- -
- Many superhuman AIs exist.
- -
- The AIs all find mutual cooperation beneficial, and have some mechanism for peaceful conflict resolution.
- -
- The AIs have diversity of final goals, so some goals are more beneficial to humans than others. This protects against any critical mistake in defining a final goal, as many goals exist. However, it is not optimal, as some of AIs may have goals that are detrimental for humans. It will be similar to our current world, with different countries, but the main difference will be that they will likely be much better able to peacefully coexist than currently, because of AI support.
- -
- Because Earth is surrounded by infinite space, different AIs could start to travel to the stars in different directions, and as each direction includes a very large number of stars, even very ambitious goals could be not mutually exclusive and might not provoke conflicts and wars.
- -
- Finding it mutually beneficial to create AI police to prevent unlimited self-improving AIs or other dangerous AIs from developing via ubiquitous intelligent control.
5.1.2. The Importance of Number in the Net of AIs
- -
- Two AI-sovereigns’ semi-stable solution, similar to the Cold War [76].
- -
- From several to dozens of sovereign AIs, similar to existing nation-states; they may be evolved from nation-states, or from large companies.
- -
- -
- Uncountable or almost-infinite number of AIs, similar to AI-medium, discussed above. This could be similar to the IoT, but with AIs as nodes.
- Net of AIs forms a multilevel immune system to protect against rogue AIs and has a diversity of values, thus including human-positive values
- ○
- Instruments to increase the number and diversity of AIs:
- -
- openness
- -
- slowdown of AI growth
- -
- human augmentation
- -
- self-improving organizations
- -
- increase number of AI teams
- -
- create many copies of the first AIs
- ○
- Net of AIs is based on human uploaded minds
- Several AI-sovereigns coexist, and they have better defensive than offensive capabilities
- ○
- Two AIs semi-stable “Cold war” solution, characterized by
- -
- tight arms race
- -
- military AI evolution
- -
- MAD defense posture
- ○
- AI-sovereigns appear from nation-states
- -
- very slow takeoff and integration with governments
- ○
- Different AIs expand in space in different directions without conflict
- ○
- Creation of AIs on remote planets
5.2. From the Arms Race between AI-Creating Teams to the Net of AIs
5.2.1. Openness in AI Development
- -
- None of the AI teams gains strategic advantage over other teams, as all the data from every team’s results are available to all of the teams. An attempt to hide results will be seen publicly. Openness ensures that many AI teams will come close to self-improving AI simultaneously, and that there will be many such AIs, which will balance each other.
- -
- The teams outside the “open net” are much less likely to gain strategic advantage, as they are not getting all the benefits of the membership in the net, namely access to the results and capabilities of others. However, this depends on how much information becomes part of the public domain. The open net will have an “intelligence advantage” over any smaller player, which makes it more probable that self-improvement will start inside the open net, or that the net will have time to react before a rogue agent “outsmarts” the net.
- -
- The “open net” will create many different AIs, which will balance each other, and probably will be motivated to engage in mutually useful collaboration (However, some could take advantage of openness of others but not share their own data and ideas). If one AI leaves the net for uncontrolled self-improvement, the collective intelligence of the net will still be higher than that AI for some time, probably enough to stop the rogue AI.
- -
- The value system of the net will provide necessary diversity, so many possible goals will be presented to at least some extent. This lessens the chance that any good goal will lost, but raises the chance that some AI projects will have bad, dangerous, or otherwise unacceptable goals.
- -
- Because of their ability to collaborate, the “open net” may be able to come to unanimous decisions about important topics, thus effectively forming a Singleton.
- -
- The net will be able to assess and possibly control all the low-hanging fruits of self-improvement, for example, the ability to buy hardware or take over the internet, thus slowing down self-improvement of any rogue agent.
- -
- The net will help to observe what the other players, not involved in open net, are doing—for example, the fact that German scientists stopped publishing articles about uranium in 1939 showed that they were trying to keep their work secret and therefore indirectly hinted that they were working on a bomb.
- -
- The net will contribute to the creation AI vigilantes or AI police, as suggested by David Brin in his transparent society proposal [73]. Therefore, the open net may somehow evolve in the direction of an AI Nanny, perhaps consisting of many distributed nodes.
5.2.2. Increase of the Number of AI Labs, so Many AIs Will Appear Simultaneously
5.2.3. Change of the Self-Improving Curve Form, So That the Distance Between Self-Improving AIs Will Diminish
5.3. Instruments to Make the Net of AIs Safer
5.3.1. Selling Cheap and Safe “Robotic Brains” Based on Non-Self-Improving Human-Like AI
5.3.2. Starting Many AIs Simultaneously
6. Solutions in Which Humans are Part of the AI System
6.1. Different Ways to Incorporate Humans inside AI
- -
- AI could be built around a human core or as a human emulation. It could result from effective personal self-improvement via neural implants [2], adding tool AIs and exocortex. There is no problem of “AI alignment,” as there are not two agents that should be aligned, but only one agent whose value system is evolving [53], however, if the human core is not aligned with the rest of humanity, the same misalignment problem could appear—therefore the ethics of the core human are crucial.
- -
- -
- AI could result from genetic modification of humans for intelligence improvement [78].
- -
- -
- Only one human upload is created, and it works as an AI Nanny, preventing the emergence of any other superintelligences [53].
- -
- Superintelligence is created by a “self-improving organization” as a property of the whole organization, which includes employees, owners, computers, hardware-building capabilities, social mechanisms, and owners. It could be a net of self-improving organizations, similar to Open AI [29] or the “Partnership on AI”.
- -
- Nation-states evolve into AI-states, and keep most of their legislation, structure, values, people, and territories. This is most probable in the case of the soft takeoff scenarios, which would take years. Earth could evolve into a bipolar world, similar to the Cold War, or a multipolar world. In this scenario, we could expect a merger between self-improving organizations and AI-states, perhaps by acquisition of such companies by state players.
6.2. Even Unfriendly AI Will Preserve Some Humans or Information about Humans
- -
- Unfriendly AI may have a subgoal to behave as benevolent AI toward humans, based on some Pascal mugging-style considerations and ontological uncertainty if it will think that there is small chance that it is in a simulation which tests its behavior [71].
- -
- Even unaligned AI will probably model humans in instrumental simulations [90] needed to solve the Fermi paradox.
- -
- Humans could be cost-effective workers in some domains and might therefore be retained, though only to be treated as slaves.
- -
- AI could preserve some humans as a potentially valuable asset, perhaps to trade information about them with potential alien AI [75], or to sell them to a benevolent AI.
- -
- AI may still preserve information about human history and DNA for billions of years, even if the AI does not use or simulate humans in the near term. It may later return them to life if it needs humans for some instrumental goal.
- -
- AI may use human “wetware” (biological brains) as efficient supercomputers.
- -
- AI could ignore humans and choose to live in space, while humans would survive on Earth. AI would preserve humanity if the marginal utility derivable from humanity’s atoms is less than the marginal instrumental utility from humanity’s continued existence.
7. Which Local Solutions Are the Best to Get a Stable Global Solution?
- (1)
- Comprehensive AI Services, which could become a basis for a system of ubiquitous surveillance and AI Police, preventing appearance of rogue AIs.
- (2)
- Research in human uploads or human-mind models, which will result in many AIs of relatively limited capabilities [74]. This again could be used to create AI Police.
- (3)
- (4)
- Robotic mind-bricks, that are pre-trained AI with limited capabilities and prefabricated safety measures which would be sold widely and provide a basis for global AI policing.
- (5)
- AI Safety as a service, similar in some sense to current antivirus computer industry.
8. Conclusions
9. Disclaimer
Author Contributions
Funding
Acknowledgments
Conflicts of Interest
References
- Yampolsky, R.; Fox, J. Safety engineering for artificial general intelligence. Topoi 2013, 32, 217–226. [Google Scholar] [CrossRef]
- Bostrom, N. Superintelligence; Oxford University Press: Oxford, UK, 2014. [Google Scholar]
- Russell, S. 3 Principles for Creating Safer AI. Available online: https://www.youtube.com/watch?v=EBK-a94IFHY (accessed on 18 February 2019).
- Sotala, K.; Yampolskiy, R. Responses to catastrophic AGI risk: A survey. Phys. Scr. 2015, 90, 069501. [Google Scholar] [CrossRef]
- Yudkowsky, E. Artificial Intelligence as a Positive and Negative Factor in Global Risk, in Global Catastrophic Risks; Cirkovic, M.M., Bostrom, N., Eds.; Oxford University Press: Oxford, UK, 2008. [Google Scholar]
- Christiano, P. Takeoff Speeds. Available online: https://sideways-view.com/2018/02/24/takeoff-speeds/ (accessed on 5 March 2018).
- Ramamoorthy, A.; Yampolskiy, R. Beyond MAD?: The race for artificial general intelligence. ICT Discov. Spec. Issue 2018, 1, 1–8. [Google Scholar]
- Brundage, M.; Avin, S.; Clark, J.; Toner, H.; Eckersley, P.; Garfinkel, B.; Dafoe, A.; Scharre, P.; Zeitzoff, T.; Filar, B. The Malicious Use of Artificial Intelligence: Forecasting, Prevention, and Mitigation. arXiv, 2018; arXiv:1802.07228. [Google Scholar]
- Sotala, K. Disjunctive AI Scenarios: Individual or Collective Takeoff? 2017. Available online: https://kajsotala.fi/2017/01/disjunctive-ai-scenarios-individual-or-collective-takeoff/ (accessed on 18 February 2019).
- Dewey, D. Long-Term Strategies for Ending Existential Risk from Fast Takeoff; Taylor & Francis: New York, NY, USA, 2016. [Google Scholar]
- Bostrom, N. What is a singleton. Linguist. Philos. Investig. 2006, 5, 48–54. [Google Scholar]
- Krakovna, V. Risks from general artificial intelligence without an intelligence explosion. Deep Saf. 2015, 26, 1–8. [Google Scholar]
- Turchin, A.; Denkenberger, D. Classification of Global Catastrophic Risks Connected with Artificial intelligence. J. Br. Interpanet. Soc. 2018, 71, 71–79. [Google Scholar] [CrossRef]
- Drexler, K.E. Reframing Superintelligence. Available online: https://www.fhi.ox.ac.uk/wp-content/uploads/Reframing_Superintelligence_FHI-TR-2019-1.1-1.pdf (accessed on 18 February 2019).
- Turchin, A. Assessing the future plausibility of catastrophically dangerous AI. Futures 2018. [Google Scholar] [CrossRef]
- Beckstead, N. On the Overwhelming Importance of Shaping the Far Future; Department of Philosophy, Rutgers University: New Brunswick, NJ, USA, 2013. [Google Scholar]
- Bostrom, N. Existential risks: Analyzing Human Extinction Scenarios and Related Hazards. J. Evol. Technol. 2002, 9, 2002. [Google Scholar]
- Torres, P. Problems with Defining an Existential Risk. Available online: https://ieet.org/index.php/IEET2/more/torres20150121 (accessed on 18 February 2019).
- Green, B.P. The Technology of Holiness: A Response to Hava Tirosh-Samuelson. Theol. Sci. 2018, 16, 223–228. [Google Scholar] [CrossRef]
- Tomasik, B. Artificial Intelligence and Its Implications for Future Suffering; Foundational Research Institute: Basel, Switzerland, 2017. [Google Scholar]
- Goertzel, B. Should Humanity Build a Global AI Nanny to Delay the Singularity Until It’s Better Understood? J. Conscious. Stud. 2012, 19, 96–111. [Google Scholar]
- Yudkowsky, E. Coherent Extrapolated Volition. Available online: http://intelligence.org/files/CEV.pdf (accessed on 18 February 2019).
- Weng, Y.-H.; Chen, C.-H.; Sun, C.-T. Safety Intelligence and Legal Machine Language: Do We Need the Three Laws of Robotics. In Service Robot Applications; InTech: Rijeka, Croatia, 2008. [Google Scholar]
- Hughes, J. Relinquishment or Regulation: Dealing with Apocalyptic Technological Threats. Hartford CT Novemb. 2001, 14, 06106. [Google Scholar]
- Yudkowsky, E. There’s No Fire Alarm for Artificial General Intelligence; Machine Intelligence Research Institute: Berkeley, CA, USA, 2017. [Google Scholar]
- Robots: Legal Affairs Committee Calls for EU-Wide Rules. Available online: http://www.europarl.europa.eu/news/en/press-room/20170110IPR57613/robots-legal-affairs-committee-calls-for-eu-wide-rules (accessed on 18 February 2019).
- Future of Life Institute Asilomar AI Principles. Available online: https://futureoflife.org/ai-principles/ (accessed on 18 February 2019).
- Morris, D.Z. Elon Musk: Artificial Intelligence Is the “Greatest Risk We Face as a Civilization”. Available online: http://fortune.com/2017/07/15/elon-musk-artificial-intelligence-2/ (accessed on 18 July 2017).
- Brockman, G.; Sutskever, I. Introducing OpenAI. Available online: https://openai.com/blog/introducing-openai/ (accessed on 18 February 2019).
- Berglas, A. Artificial intelligence will kill our grandchildren (singularity). Unpublished work. 2012. [Google Scholar]
- Green, B. Are science, technology, and engineering now the most important subjects for ethics? Our need to respond. In Proceedings of the 2014 IEEE International Symposium on Ethics in Science, Technology and Engineering, Chicago, IL, USA, 23–24 May 2014; pp. 1–7. [Google Scholar]
- Green, B. Emerging technologies, catastrophic risks, and ethics: three strategies for reducing risk. In Proceedings of the 2016 IEEE International Symposium on Ethics in Engineering, Science and Technology (ETHICS), Vancouver, BC, Canada, 13–14 May 2016. [Google Scholar]
- List of Semiconductor Fabrication Plants. Available online: https://en.wikipedia.org/wiki/List_of_semiconductor_fabrication_plants (accessed on 18 February 2019).
- Cole, D.D.; Denkenberger, D.; Griswold, M.; Abdelkhaliq, M.; Pearce, J. Feeding Everyone if Industry is Disabled. In Proceedings of the 6th International Disaster and Risk Conference, Davos, Switzerland, 28 August–1 September 2016. [Google Scholar]
- Denkenberger, D.; Cole, D.; Griswold, M.; Pearce, J.; Taylor, A.R. Non Food Needs if Industry is Disabled. In Proceedings of the Proceedings of the 6th International Disaster and Risk Conference, Davos, Switzerland, 28 August–1 September 2016. [Google Scholar]
- Jones, S.E. Against Technology: From the Luddites to Neo-Luddism; Routledge: Abingdon, UK, 2013; ISBN 1-135-52239-1. [Google Scholar]
- Kushner, D. The real story of stuxnet. IEEE Spectr. 2013, 50, 48–53. [Google Scholar] [CrossRef]
- Bostrom, N. The Unilateralist’s Curse: The Case for a Principle of Conformity. Available online: http://www.nickbostrom.com/papers/unilateralist.pdf (accessed on 18 February 2019).
- Tegmark, M. Life 3.0: Being Human in the Age of Artificial Intelligence; Knopf: New York, NY, USA, 2017. [Google Scholar]
- Turchin, A.; Denkenberger, D. Military AI as convergent goal of the self-improving AI. In Artificial Intelligence Safety and Security; CRC Press: Louisville, KY, USA, 2018. [Google Scholar]
- Teller, E. LA-602: The Ignition of Atmosphere with Nuclear Bombs; Los Alamos Laboratory: Los Alamos, NM, USA, 1946.
- Ria Novosti Испытания ядернoгo oружия на Тoцкoм пoлигoне. Справка. Available online: https://ria.ru/defense_safety/20090914/184923659.html (accessed on 18 July 2017).
- Nuclearweaponarchive India’s Nuclear Weapons Program—Smiling Buddha: 1974. Available online: http://nuclearweaponarchive.org/India/IndiaSmiling.html (accessed on 18 July 2017).
- MIRI. MIRI AMA—Anyone May Ask. Available online: http://effective-altruism.com/r/main/ea/12r/ask_miri_anything_ama/ (accessed on 20 February 2019).
- MIRI. About MIRI. Available online: https://intelligence.org/about/ (accessed on 18 February 2019).
- Sotala, K. Decisive Strategic Advantage without a Hard Takeoff. 2016. Available online: https://kajsotala.fi/2016/04/decisive-strategic-advantage-without-a-hard-takeoff/ (accessed on 18 February 2019).
- Putin, V. Open Lesson “Russia Looking to the Future”. Available online: http://kremlin.ru/events/president/news/55493 (accessed on 28 October 2017).
- Kahn, H. On Thermonuclear War; Princeton University Press: Princeton, NJ, USA, 1959. [Google Scholar]
- Muehlhauser, L.; Salamon, A. Intelligence Explosion: Evidence and Import. In Singularity Hypotheses; Springer: Berlin/Heidelberg, Germany, 2012. [Google Scholar]
- Christiano, P. Prosaic AI Alignment. Available online: https://ai-alignment.com/prosaic-ai-control-b959644d79c2 (accessed on 18 February 2019).
- Itut Reality Check: ‘We Are Not Nearly As Close To Strong AI As Many Believe’. Available online: https://news.itu.int/reality-check-not-nearly-close-strong-ai-many-believe/ (accessed on 18 February 2019).
- Partnership for AI. Available online: https://www.partnershiponai.org/ (accessed on 18 February 2019).
- Turchin, A. Human Upload as AI Nanny 2017. Available online: https://www.academia.edu/38386976/Human_upload_as_AI_Nanny (accessed on 19 February 2019).
- Mahoney, M. A Proposed Design for Distributed Artificial General Intelligence. 2008. Available online: http://mattmahoney.net/agi2.html (accessed on 18 February 2019).
- Alexander, S. Ascended Economy? Available online: http://slatestarcodex.com/2016/05/30/ascended-economy/ (accessed on 18 February 2019).
- Hanson, R.; Sun, W. Probability and Asset Updating using Bayesian Networks for Combinatorial Prediction Markets. arXiv, 2012; arXiv:1210.4900. [Google Scholar]
- Camarinha-Matos, L.M.; Afsarmanesh, H. Collaborative networks: a new scientific discipline. J. Intell. Manuf. 2005, 16, 439–452. [Google Scholar] [CrossRef]
- Luksha, P. NeuroWeb Roadmap: Results of Foresight & Call for Action. 2014. Available online: https://dlib.si/details/URN:NBN:SI:DOC-IXKS9ZQW (accessed on 18 February 2019).
- Vinding, M. Reflections on Intelligence; Heinemann: Portsmouth, NJ, USA, 2016. [Google Scholar]
- Marx, K. Capital: A Critique of Political Economy. The Process of Production of Capital. 1867. Available online: https://oll.libertyfund.org/titles/marx-capital-a-critique-of-political-economy-volume-i-the-process-of-capitalist-production (accessed on 18 February 2019).
- Smart, J. The transcension hypothesis: Sufficiently advanced civilizations invariably leave our universe, and implications for METI and SETI. Acta Astronaut. 2012, 78, 55–68. [Google Scholar] [CrossRef] [Green Version]
- Gwern. Why Tool AIs want to be Agent AIs 2016. Available online: https://www.gwern.net/Tool-AI (accessed on 18 February 2019).
- Yudkowsky, E. Harry Potter and Method of Rationality. 2010. Available online: https://fanlore.org/wiki/Harry_Potter_and_the_Methods_of_Rationality (accessed on 18 February 2019).
- Bostrom, N.; Armstrong, S.; Shulman, C. Racing to the Precipice: a Model of Artificial Intelligence Development. AI Soc. 2013, 31, 201–206. [Google Scholar]
- Shulman, C. Arms races and intelligence explosions. In Singularity Hypotheses; Springer: Berlin, Germany, 2011. [Google Scholar]
- Bostrom, N. Strategic Implications of Openness in AI Development. Glob. Policy 2016, 8, 135–148. [Google Scholar] [CrossRef]
- Baum, S.D. On the Promotion of Safe and Socially Beneficial Artificial Intelligence. Glob. Catastroph. Risk. 2016, 32, 543–551. [Google Scholar] [CrossRef]
- Ouagrham-Gormley, S.B. Dissuading Biological Weapons. In Proliferation Pages; Springer: Berlin, Germany, 2013; pp. 473–500. [Google Scholar]
- Auerbach, D. The Most Terrifying Thought Experiment of All Time. Available online: http://www.slate.com/articles/technology/bitwise/2014/07/roko_s_basilisk_the_most_terrifying_thought_experiment_of_all_time.html (accessed on 18 February 2019).
- Fernando, C. PathNet: Evolution Channels Gradient Descent in Super Neural Networks. arXiv, 2017; arXiv:1701.08734. [Google Scholar]
- Nelson, R. How to Deter a Rogue AI by Using Your First-mover Advantage. 2007. Available online: http://www.sl4.org/archive/0708/16600.html (accessed on 18 February 2019).
- Kharpal, A. Elon Musk: Humans Must Merge with Machines or Become Irrelevant in AI Age; CNBC: Englewood Cliffs, NJ, USA, 2017. [Google Scholar]
- Brin, D. The Transparent Society; Perseus Book: New York, NY, USA, 1998. [Google Scholar]
- Hanson, R. The Age of Em: Work, Love, and Life when Robots Rule the Earth; Oxford University Press: Oxford, UK, 2016. [Google Scholar]
- Bostrom, N. Hail Mary, Value Porosity, and Utility Diversification; Oxford University Press: Oxford, UK, 2016. [Google Scholar]
- Lem, S. The Investigation. 1959. Available online: https://onlinelibrary.wiley.com/doi/abs/10.1111/j.1651-2227.1959.tb05423.x (accessed on 18 February 2019).
- Urban, T. Neuralink and the Brain’s Magical Future. 2017. Available online: https://waitbutwhy.com/2017/04/neuralink.html (accessed on 18 February 2019).
- Bostrom, N. Human genetic enhancements: a transhumanist perspective. J. Value Inq. 2003, 37, 493–506. [Google Scholar] [CrossRef]
- Smith, P.D. Doomsday Men: The Real Dr. Strangelove and the Dream of the Superweapon; St. Martin’s Press: New York, NY, USA, 2007. [Google Scholar]
- Alexander, S. Should AI Be Open. Available online: https://slatestarcodex.com/2015/12/17/should-ai-be-open/ (accessed on 18 February 2019).
- Baker, B.H. The Gray Matter: The Forgotten Story of the Telephone; Telepress: Kent, WA, USA, 2000; ISBN 0-615-11329-X. [Google Scholar]
- The Telegraph Russian Spacecraft Landed on Moon Hours Before Americans. Available online: http://www.telegraph.co.uk:80/science/space/5737854/Russian-spacecraft-landed-on-moon-hours-before-Americans.html (accessed on 18 February 2019).
- Venture Scanner Artificial Intelligence Q1 Update in 15 Visuals 2016. Available online: https://www.venturescanner.com/blog/2016/artificial-intelligence-q1-update-in-15-visuals (accessed on 18 February 2019).
- Yampolskiy, R. From Seed AI to Technological Singularity via Recursively Self-Improving Software. arXiv, 2015; arXiv:1502.06512. [Google Scholar]
- Drexler, E.; Phoenix, C. Safe exponential manufacturing. Nanotechnology 2004, 15, 869. [Google Scholar]
- Bontchev, V. Are Good Computer Viruses Still a Bad Idea? EICAR: London, UK, 1994. [Google Scholar]
- Sotala, K.; Valpola, H. Coalescing minds: brain uploading-related group mind scenarios. Int. J. Mach. Conscious. 2012, 4, 293–312. [Google Scholar] [CrossRef]
- Batin, M.; Turchin, A.; Markov, S.; Zhila, A.; Denkenberger, D. Artificial Intelligence in Life Extension: From Deep Learning to Superintelligence. Inform. Slov. 2018, 41, 401. [Google Scholar]
- Alexander, S. Book Review: Age of Em. Available online: http://slatestarcodex.com/2016/05/28/book-review-age-of-em/ (accessed on 18 February 2019).
- Bostrom, N. Are You Living in a Computer Simulation? Publ. Philos. Q. 2003, 53, 243–255. [Google Scholar] [CrossRef]
- Omohundro, S. The basic AI drives. In Proceedings of the AGI Frontiers in Artificial Intelligence and Applications, Memphis, TN, USA, 1–3 March 2008. [Google Scholar]
- Bostrom, N. Existential risk prevention as global priority. Glob. Policy 2013, 4, 15–31. [Google Scholar] [CrossRef]
- Shakirov, V. Review of State-of-the-Arts in Artificial Intelligence with Application to AI Safety Problem. arXiv, 2016; arXiv:1605.04232. [Google Scholar]
- DeepMind AlphaGo. Available online: https://deepmind.com/research/alphago/ (accessed on 18 February 2019).
- Ministry of National Defense of the People’s Republic of China. The Dawn of the Intelligent Military Revolution; Ministry of National Defense of the People’s Republic of China: Beijing, China, 2016.
- Factored Cognition (May 2018) Ought. Available online: https://ought.org/presentations/factored-cognition-2018-05 (accessed on 25 January 2019).
© 2019 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).
Share and Cite
Turchin, A.; Denkenberger, D.; Green, B.P. Global Solutions vs. Local Solutions for the AI Safety Problem. Big Data Cogn. Comput. 2019, 3, 16. https://doi.org/10.3390/bdcc3010016
Turchin A, Denkenberger D, Green BP. Global Solutions vs. Local Solutions for the AI Safety Problem. Big Data and Cognitive Computing. 2019; 3(1):16. https://doi.org/10.3390/bdcc3010016
Chicago/Turabian StyleTurchin, Alexey, David Denkenberger, and Brian Patrick Green. 2019. "Global Solutions vs. Local Solutions for the AI Safety Problem" Big Data and Cognitive Computing 3, no. 1: 16. https://doi.org/10.3390/bdcc3010016
APA StyleTurchin, A., Denkenberger, D., & Green, B. P. (2019). Global Solutions vs. Local Solutions for the AI Safety Problem. Big Data and Cognitive Computing, 3(1), 16. https://doi.org/10.3390/bdcc3010016