{"id":2318,"date":"2021-01-12T00:00:00","date_gmt":"2021-01-12T00:00:00","guid":{"rendered":"http:\/\/localhost\/wptest\/2023\/09\/06\/advanced-machine-intelligence-orthogonality-instrumental-convergence-and-the-dangers-of-value-misalignment\/"},"modified":"2021-01-12T00:00:00","modified_gmt":"2021-01-12T00:00:00","slug":"advanced-machine-intelligence-orthogonality-instrumental-convergence-and-the-dangers-of-value-misalignment-2","status":"publish","type":"post","link":"https:\/\/blogs.ncl.ac.uk\/philosophy\/2021\/01\/12\/advanced-machine-intelligence-orthogonality-instrumental-convergence-and-the-dangers-of-value-misalignment-2\/","title":{"rendered":"Advanced Machine Intelligence: Orthogonality, Instrumental Convergence and the Dangers of Value Misalignment"},"content":{"rendered":"<p>The object of this dissertation is artificial intelligence (AI), and in particular it concerns AI risk or AI safety. I argue for the veracity of Bostrom\u2019s orthogonality thesis (2012) \u2013 contextualised with reference to Hume\u2019s (2007) is-ought distinction \u2013 and instrumental convergence thesis (developed initially by Omohundro (2008) in terms of the \u201cBasic AI Drives\u201d). In combination, what these theses show is that the default outcome of advanced AI (AGI and ASI) is existential catastrophe, and thereby the importance of ensuring that the value systems of advanced artificial agents are human compatible. I consider two main approaches to the value alignment problem \u2013 direct specification and value learning \u2013 and point out the flaws in each. While this project does not offer its own approach value alignment, the central concern of AI safety, it does emphasise the necessity for AI research to undergo a perspectival shift and focus on the search for one. The AI community should, that is, be concerned foremost with AI safety rather than AI capability.<\/p>\n","protected":false},"excerpt":{"rendered":"<p>Edward Potts, 2021, Stage 3<\/p>\n","protected":false},"author":8792,"featured_media":0,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[311,22,123],"tags":[148,270,69],"class_list":["post-2318","post","type-post","status-publish","format-standard","hentry","category-311","category-abstracts","category-stage-3-abstracts","tag-artificial-intelligence","tag-humanity","tag-technology"],"_links":{"self":[{"href":"https:\/\/blogs.ncl.ac.uk\/philosophy\/wp-json\/wp\/v2\/posts\/2318","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/blogs.ncl.ac.uk\/philosophy\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/blogs.ncl.ac.uk\/philosophy\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/blogs.ncl.ac.uk\/philosophy\/wp-json\/wp\/v2\/users\/8792"}],"replies":[{"embeddable":true,"href":"https:\/\/blogs.ncl.ac.uk\/philosophy\/wp-json\/wp\/v2\/comments?post=2318"}],"version-history":[{"count":0,"href":"https:\/\/blogs.ncl.ac.uk\/philosophy\/wp-json\/wp\/v2\/posts\/2318\/revisions"}],"wp:attachment":[{"href":"https:\/\/blogs.ncl.ac.uk\/philosophy\/wp-json\/wp\/v2\/media?parent=2318"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/blogs.ncl.ac.uk\/philosophy\/wp-json\/wp\/v2\/categories?post=2318"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/blogs.ncl.ac.uk\/philosophy\/wp-json\/wp\/v2\/tags?post=2318"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}