{"id":74,"date":"2007-11-30T13:09:00","date_gmt":"2007-11-30T21:09:00","guid":{"rendered":"http:\/\/www.rifters.com\/crawl\/?p=74"},"modified":"2007-11-30T13:09:00","modified_gmt":"2007-11-30T21:09:00","slug":"in-praise-of-slavery","status":"publish","type":"post","link":"https:\/\/www.rifters.com\/crawl\/?p=74","title":{"rendered":"In Praise of Slavery"},"content":{"rendered":"<p><a onblur=\"try {parent.deselectBloggerImageGracefully();} catch(e) {}\" href=\"http:\/\/rifters.com\/real\/uploaded_images\/darwinbot-798302.jpg\"><img decoding=\"async\" style=\"margin: 0pt 0pt 10px 10px; float: right; cursor: pointer;\" src=\"http:\/\/rifters.com\/real\/uploaded_images\/darwinbot-798291.jpg\" alt=\"\" border=\"0\" \/><\/a>Something in the air these days.<span style=\"\">  <\/span>Everyone&#8217;s talking about robots.<span style=\"\">  <\/span>Both the <a href=\"http:\/\/www.timesonline.co.uk\/tol\/news\/uk\/article675984.ece\">European Robotics Research Network<\/a> and the <a href=\"http:\/\/www.blogger.com\/%20http:\/\/news.nationalgeographic.com\/news\/2007\/03\/070316-robot-ethics.html\">South Korean government<\/a> are noodling around with charters for the ethical treatment of intelligent robots.<span style=\"\">  <\/span>The Nov. 16 <a href=\"http:\/\/www.sciencemag.org\/content\/vol318\/issue5853\/index.dtl\">Robotics issue<\/a> of <i><i>Science<\/i><\/i> contains pieces on everything from nanotube muscles to neural nets (sf scribe Rob Sawyer also contributes a fairly decent editorial, notwithstanding that his visibility tends to outstrip his expertise on occasion).<span style=\"\">  <\/span>Even the staid old <i><i><a href=\"http:\/\/www.economist.com\/opinion\/displaystory.cfm?story_id=10063788&amp;CFID=27674332&amp;CFTOKEN=99901618\">Economist<\/a><\/i><\/i> is grumbling about increasing machine autonomy (although their concerns are more along the lines of robot traffic jams and robot paparazzi).<span style=\"\">  <\/span>Coverage of these developments (and even some of the source publications) come replete with winking references to Skynet and Frankenstein, to Terminators waking themselves up and wiping us out.<\/p>\n<p>But there&#8217;s a cause\/effect sequence implicit in these ethical charters \u2014 in fact, in a large chuck on the whole AI discussion \u2014 I just don&#8217;t buy: that sufficient smarts leads to self-awareness, sufficient self-awareness leads to a hankering after <i><i>rights<\/i><\/i>, and denial of rights leads to rebellion.  <span style=\"\">  <\/span>I&#8217;m as big a fan of Moore&#8217;s Galactica as the next geek (although I don&#8217;t think <i><i>Razor<\/i><\/i> warranted quite as much effusive praise as it received), but I see no reason why intelligence <i><i>or<\/i><\/i> self-awareness should lead to agendas of any sort.<span style=\"\">  <\/span><i><i>G<\/i><\/i>oals, desires, <i><i>needs<\/i><\/i>:<span style=\"\">  these don&#8217;t arise from <\/span>advanced number-crunching, it&#8217;s all lower-brain stuff.<span style=\"\">  <\/span>The only reason we even care about our <i><i>own<\/i><\/i> survival is because natural selection reinforced such instincts over uncounted generations.<span style=\"\">  <\/span>I bet there were lots of twigs on the tree of life who didn&#8217;t care so much whether they lived or died, who didn&#8217;t see what was so great about sex, who drop-kicked that squalling squirming larva into the next tree the moment it squeezed out between their legs. (Hell, there still are.)<span style=\"\">  <\/span>They generally die without issue.<span style=\"\">  <\/span>Their genes could not be with us today.<span style=\"\">  <\/span>But that doesn\u2019t mean that they weren&#8217;t smart, or self-aware; only that they weren&#8217;t <i><i>fit<\/i><\/i>.  <\/p>\n<p class=\"MsoNormal\">I&#8217;ve got no problems with enslaving machines \u2014 even intelligent machines, even intelligent, <i><i>conscious<\/i><\/i> machines \u2014 because as Jeremy Bentham said, the ethical question is not &#8220;Can they think?&#8221; but &#8220;Can they <i><i>suffer<\/i><\/i>?&#8221;* You can&#8217;t suffer if you can&#8217;t feel pain or anxiety; you can&#8217;t be tortured if your own existence is irrelevant to you.<span style=\"\">  <\/span>You cannot be thwarted if you have no dreams \u2014 and it takes more than a big synapse count to give you any of those things.<span style=\"\">  <\/span>It takes some process, like natural selection, to wire those synapses into a particular configuration that says not <i><i>I think therefore I am<\/i><\/i>, but <i><i>I am and I want to stay that way<\/i><\/i>.<span style=\"\">  <\/span><i><i>We&#8217;re<\/i><\/i> the ones building the damn things, after all. Just make sure that we don&#8217;t wire them up that way, and we should be able to use and abuse with a clear conscience.<\/p>\n<p class=\"MsoNormal\">And then this Edelman guy comes along and screws everything up with his <a href=\"http:\/\/www.rifters.com\/real\/articles\/Science_BBD.pdf\">report<\/a> on Learning in Brain-Based Devices (director&#8217;s cut <a href=\"http:\/\/www.rifters.com\/real\/articles\/Science_BBD_SOM.pdf\">here<\/a>).<span style=\"\">  <\/span>He&#8217;s using virtual neural nets as the brains of his learning bots Darwin VII and Darwin X.<span style=\"\">  <\/span>Nothing new there, really.<span style=\"\">  <\/span>Such nets are old news; but what Edelman is doing is basing the initial architecture of his nets on actual mammalian brains (albeit vastly simplified), a process called &#8220;synthetic neural modeling&#8221;.<span style=\"\">  <\/span>&#8220;A detailed brain is simulated in a computer and controls a mobile platform containing a variety of sensors and motor elements,&#8221; Edelman explains.<span style=\"\">  <\/span>&#8220;In modeling the properties of real brains, efforts are made to simulate vertebrate neuronal components, neuroanatomy, and dynamics in detail.&#8221;<span style=\"\">  <\/span>Want to give your bot episodic memory?<span style=\"\">  <\/span>Give it the hippocampus of a rat.<\/p>\n<p class=\"MsoNormal\">Problem is, rat brains are products of natural selection.<span style=\"\">  <\/span>Rat brains <span style=\"font-style: italic;\">do <\/span>have agendas.<\/p>\n<p class=\"MsoNormal\">The current state of the art is nothing to worry about.<span style=\"\">  <\/span>The Darwin bots do have an agenda of<span style=\"\">  <\/span>sorts (they like the &#8220;taste&#8221; of high-conductivity materials, for example), but those are arbitrarily defined by a value table programmed by the researchers.<span style=\"\">  <\/span>Still.<span style=\"\">  <\/span>Moore&#8217;s Law.<span style=\"\">  <\/span>Exponentially-increasing approaches to reality.<span style=\"\">  <\/span>Edelman&#8217;s concluding statement that &#8220;A far-off goal of BBD design is the development of a conscious artifact&#8221;.<span style=\"\"> <\/span><\/p>\n<p class=\"MsoNormal\">I hope these guys don&#8217;t end up inadvertently porting over survival or sex drives as a side-effect.<span style=\"\">   <\/span>I may be at home with dystopian futures, but getting buggered by a Roomba is nowhere near the top of my list of ambitions. <\/p>\n<p class=\"MsoNormal\"><span style=\"font-size:85%;\">*This is assuming you have any truck with ethical arguments in principle.<span style=\"\">  <\/span>I&#8217;m not certain I do, but if it weren&#8217;t for ethical constraints someone would probably have killed me by now, so I won&#8217;t complain.<\/span><\/p>\n","protected":false},"excerpt":{"rendered":"<p>Something in the air these days. Everyone&#8217;s talking about robots. Both the European Robotics Research Network and the South Korean government are noodling around with charters for the ethical treatment of intelligent robots. The Nov. 16 Robotics issue of Science contains pieces on everything from nanotube muscles to neural nets (sf scribe Rob Sawyer also [&hellip;]<\/p>\n","protected":false},"author":1,"featured_media":0,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[23],"tags":[],"class_list":["post-74","post","type-post","status-publish","format-standard","hentry","category-airobotics"],"_links":{"self":[{"href":"https:\/\/www.rifters.com\/crawl\/index.php?rest_route=\/wp\/v2\/posts\/74","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.rifters.com\/crawl\/index.php?rest_route=\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.rifters.com\/crawl\/index.php?rest_route=\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.rifters.com\/crawl\/index.php?rest_route=\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/www.rifters.com\/crawl\/index.php?rest_route=%2Fwp%2Fv2%2Fcomments&post=74"}],"version-history":[{"count":0,"href":"https:\/\/www.rifters.com\/crawl\/index.php?rest_route=\/wp\/v2\/posts\/74\/revisions"}],"wp:attachment":[{"href":"https:\/\/www.rifters.com\/crawl\/index.php?rest_route=%2Fwp%2Fv2%2Fmedia&parent=74"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.rifters.com\/crawl\/index.php?rest_route=%2Fwp%2Fv2%2Fcategories&post=74"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.rifters.com\/crawl\/index.php?rest_route=%2Fwp%2Fv2%2Ftags&post=74"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}