{"id":7580,"date":"2020-08-02T20:11:29","date_gmt":"2020-08-02T20:11:29","guid":{"rendered":"http:\/\/theoreti.ca\/?p=7580"},"modified":"2022-08-30T19:28:25","modified_gmt":"2022-08-30T19:28:25","slug":"philosophers-on-gpt-3","status":"publish","type":"post","link":"https:\/\/theoreti.ca\/?p=7580","title":{"rendered":"Philosophers On GPT-3"},"content":{"rendered":"<blockquote><p><a href=\"https:\/\/dailynous.com\/2020\/07\/30\/philosophers-gpt-3\/#chalmers\"><img decoding=\"async\" class=\"alignnone size-full\" src=\"http:\/\/theoreti.ca\/wp-content\/uploads\/2020\/08\/binary-hex-pixelated-banner-fade.png\" alt=\"\" \/><\/a><\/p>\n<p>GPT-3 raises many philosophical questions. Some are ethical. Should we develop and deploy GPT-3, given that it has many biases from its training, it may displace human workers, it can be used for deception, and it could lead to AGI? I\u2019ll focus on some issues in the philosophy of mind. Is GPT-3 really intelligent, and in what sense? Is it conscious? Is it an agent? Does it understand?<\/p><\/blockquote>\n<p>On the Daily Nous (news by and for philosophers) there is a great collection of short essays on <a href=\"https:\/\/openai.com\/\">OpenAI<\/a>&#8216;s recently released API to <a href=\"https:\/\/www.theguardian.com\/technology\/2020\/jun\/12\/elon-musk-backed-openai-to-release-text-tool-it-called-dangerous\">GPT-3<\/a>, see <a href=\"https:\/\/dailynous.com\/2020\/07\/30\/philosophers-gpt-3\/\">Philosophers On GPT-3 (updated with replies by GPT-3)<\/a>. And &#8230; there is a response from GPT-3. Some of the issues raised include:<\/p>\n<p><strong>Ethics:\u00a0<\/strong>David Chalmers raises the inevitable ethics issues. Remember that GPT-2 was considered so good as to be dangerous. I don&#8217;t know if it is brilliant marketing or genuine concern, but OpenAI is continuing to treat this technology as something to be careful about. Here is Chalmers on ethics,<\/p>\n<blockquote><p>GPT-3 raises many philosophical questions. Some are ethical. Should we develop and deploy GPT-3, given that it has many biases from its training, it may displace human workers, it can be used for deception, and it could lead to AGI? I\u2019ll focus on some issues in the philosophy of mind. Is GPT-3 really intelligent, and in what sense? Is it conscious? Is it an agent? Does it understand?<\/p><\/blockquote>\n<p>Annette Zimmerman in her essay makes an important point about the larger justice context of tools like GPT-3. It is not just a matter of ironing out the biases in the language generated (or used in training.) It is not a matter of finding a techno-fix that makes bias go away. It is about care.<\/p>\n<blockquote><p>Not all uses of AI, of course, are inherently objectionable, or automatically unjust\u2014the point is simply that much like\u00a0<a href=\"https:\/\/www.hup.harvard.edu\/catalog.php?isbn=9780674411524\">we can do things with words<\/a>, we can\u00a0<em><strong>do <\/strong><\/em>things with algorithms and machine learning models. This is not purely a tangibly material\u00a0<strong><em>distributive justice <\/em><\/strong>concern: especially in the context of language models like GPT-3, paying attention to other facets of injustice\u2014<strong><em>relational, communicative, representational, ontological<\/em><\/strong>\u2014is essential.<\/p><\/blockquote>\n<p>She also makes an important and deep point that any AI application will have to make use of concepts from the application domain and all of these concepts will be contested. There are no simple concepts just as there are no concepts that don&#8217;t change over time.<\/p>\n<p>Finally, Shannon Vallor has an essay that revisits Hubert Dreyfus&#8217;s critique of AI as not really understanding.<\/p>\n<blockquote><p>Understanding is beyond GPT-3\u2019s reach because understanding cannot occur in an isolated behavior, no matter how clever. Understanding is not an act but a\u00a0<strong><em>labor<\/em><\/strong>.<\/p><\/blockquote>\n","protected":false},"excerpt":{"rendered":"<p>GPT-3 raises many philosophical questions. Some are ethical. Should we develop and deploy GPT-3, given that it has many biases from its training, it may displace human workers, it can be used for deception, and it could lead to AGI? I\u2019ll focus on some issues in the philosophy of mind. Is GPT-3 really intelligent, and &hellip; <a href=\"https:\/\/theoreti.ca\/?p=7580\" class=\"more-link\">Continue reading <span class=\"screen-reader-text\">Philosophers On GPT-3<\/span><\/a><\/p>\n","protected":false},"author":5,"featured_media":0,"comment_status":"closed","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[74,58,27,18],"tags":[],"class_list":["post-7580","post","type-post","status-publish","format-standard","hentry","category-artificial-intelligence","category-ethics-of-data-science","category-media-and-news","category-philosophy-of-computing"],"_links":{"self":[{"href":"https:\/\/theoreti.ca\/index.php?rest_route=\/wp\/v2\/posts\/7580","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/theoreti.ca\/index.php?rest_route=\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/theoreti.ca\/index.php?rest_route=\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/theoreti.ca\/index.php?rest_route=\/wp\/v2\/users\/5"}],"replies":[{"embeddable":true,"href":"https:\/\/theoreti.ca\/index.php?rest_route=%2Fwp%2Fv2%2Fcomments&post=7580"}],"version-history":[{"count":5,"href":"https:\/\/theoreti.ca\/index.php?rest_route=\/wp\/v2\/posts\/7580\/revisions"}],"predecessor-version":[{"id":8115,"href":"https:\/\/theoreti.ca\/index.php?rest_route=\/wp\/v2\/posts\/7580\/revisions\/8115"}],"wp:attachment":[{"href":"https:\/\/theoreti.ca\/index.php?rest_route=%2Fwp%2Fv2%2Fmedia&parent=7580"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/theoreti.ca\/index.php?rest_route=%2Fwp%2Fv2%2Fcategories&post=7580"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/theoreti.ca\/index.php?rest_route=%2Fwp%2Fv2%2Ftags&post=7580"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}