{"id":4972,"date":"2015-12-24T01:52:36","date_gmt":"2015-12-24T06:52:36","guid":{"rendered":"https:\/\/afutureworththinkingabout.com\/?p=4972"},"modified":"2023-02-08T13:53:47","modified_gmt":"2023-02-08T18:53:47","slug":"2012-strange-things-happen-at-the-one-two-point-the-implications-of-aci-in-sf-media","status":"publish","type":"post","link":"https:\/\/afutureworththinkingabout.com\/?p=4972","title":{"rendered":"(2012) Strange Things Happen at the One Two Point: The Implications of ACI in SF Media"},"content":{"rendered":"<p>It&#8217;s been quite some time (three years) since it was done, and some of the <a href=\"https:\/\/twitter.com\/Wolven\/status\/677682444544380929\" target=\"_blank\" rel=\"noopener noreferrer\">recent conversations<\/a> I&#8217;ve been having about <a href=\"https:\/\/twitter.com\/Wolven\/status\/679452664338493440\" target=\"_blank\" rel=\"noopener noreferrer\">machine consciousness<\/a> reminded me that I never posted the text to my paper from the joint session of the <a href=\"http:\/\/www.iacap.org\/\">International Association for Computing And Philosophy<\/a> and the <span class=\"st\"><a href=\"http:\/\/www.aisb.org.uk\/\" target=\"_blank\" rel=\"noopener noreferrer\">The British Society for the Study of Artificial Intelligence and the Simulation of Behaviour<\/a>, back in 2012.<\/span><\/p>\n<p><span class=\"st\">That year&#8217;s joint ASIB\/IACAP session was also a celebration of <a href=\"http:\/\/www.turing.org.uk\/index.html\" target=\"_blank\" rel=\"noopener noreferrer\">Alan Turing<\/a>&#8216;s centenary, and it contained <a href=\"http:\/\/machinequestion.org\/symposium\/\" target=\"_blank\" rel=\"noopener noreferrer\">The Machine Question Symposium<\/a>, an exploration of multiple perspectives on machine intelligence ethics, put together by <a href=\"http:\/\/www.gunkelweb.com\/\" target=\"_blank\" rel=\"noopener noreferrer\">David J Gunkel<\/a> and <a href=\"http:\/\/www.cs.bath.ac.uk\/~jjb\" target=\"_blank\" rel=\"noopener noreferrer\">Joanna J Bryson<\/a>. So I modded a <a href=\"http:\/\/www.needcoffee.com\/2010\/06\/09\/splice-movie-review\/\" target=\"_blank\" rel=\"noopener noreferrer\">couple<\/a> of <a href=\"http:\/\/www.needcoffee.com\/2010\/08\/26\/sarah-connor-chronicles-strange-things-happen\/\" target=\"_blank\" rel=\"noopener noreferrer\">articles<\/a> I wrote on fictional depictions of created life for <a href=\"http:\/\/NeedCoffee.com\" target=\"_blank\" rel=\"noopener noreferrer\">NeedCoffee.com<\/a>, back in 2010, beefed up the research and citations a great deal, and was thus afforded my first (but by no means last) conference appearance requiring international travel. There are, in here, the seeds of many other posts that you&#8217;ll find on this blog.<br \/>\n<\/span><\/p>\n<p>So, below the cut, you&#8217;ll find the full text of the paper, and a picture of the poster session I presented. If you&#8217;d rather not click through, you can find both of those things <a href=\"https:\/\/www.academia.edu\/4230545\/Strange_Things_Happen_at_the_One_Two_Point_The_Implications_of_Autonomous_Created_Intelligence_in_Speculative_Fiction_Media\" target=\"_blank\" rel=\"noopener noreferrer\">at this link<\/a>.<\/p>\n<p><!--more--><\/p>\n<p><strong>ABSTRACT: <\/strong>By its very nature, Science Fiction media has often concerned itself with advances in human enhancement as well as the creation of various autonomous, thinking, non-human beings. Unfortunately, since the initial proffering of the majority interpretation of Frankenstein, Mary Shelly&#8217;s seminal work, and before, most speculative fiction media has taken the standpoint that to enhance or to explore the creation of intelligences, in this way, is doomed to failure, thus recapitulating the myths of Daedalus and of Prometheus and of Lucifer, again and again. What we see and are made to fear are the uprisings of the robots or the artificial neural networks, rather than discussing and respecting the opportunity for a non-human intelligence to arise and demand rights.<\/p>\n<p>In this work, I make use of specific films, books, and television shows to explore the philosophical and cultural implications of an alternate interpretation of not only Frankenstein, but of the whole of the field of science fiction. In the first part I argue that it isn&#8217;t humanity&#8217;s attempts to \u201cplay god\u201d that cause our failures, but rather our refusal or inability to pay attention to our circumstances, to take responsibility for our creations, and to learn from the warnings and mistakes of those who went before us. Only with this recognition in mind can we move on to accepting and respecting the <em>fundamental otherness<\/em> of the intelligences we may either create or cause to be created, all while seeking to <em>bridge<\/em> that otherness, and come to mutual understanding.<\/p>\n<p>As humans have evolved, their concerns have become those of biological creatures with biologically directed needs. Food, shelter, emotional comfort, and stability are needs which would not necessarily occur to an intelligence without the organic component. It would therefore fall to humans to A) Initially recognise the concerns of such an intelligence; B) Countenance and <em>concretise<\/em> said concerns, in the understanding of other humans; and C) Create a system of interaction through which human concerns were conveyed to these new intelligences, not as <em>primary<\/em>, but as co-equal. We will do this only by considering that which causes our assumptions and cultural behaviour, namely the stories which we engage, as a culture, and deconstructing both their content and their impact.<\/p>\n<p>In all fictional considerations of non-human, and specifically machine intelligence, there is an element of fear of that which we have created. This horror at being \u201creplaced\u201d or \u201cmade obsolete\u201d drives us to regard robots and autonomous created intelligences as nothing more than tools to be used, an operational mode which leads to the assumption that machines cannot have rights or even be considered as conscious minds. This assumption begs the question, in the extreme. It is my contention that, with a proper formulation of the rights and responsibilities of and to both human and non-human persons\u2014with consideration for the necessary variance of concerns within different compositions of intelligences\u2014an understanding may be reached wherein our future societies account for not only human needs and development, but those of <em>all<\/em> intelligences, whatever form they may take.<\/p>\n<h3 style=\"text-align: center;\"><strong>1 INTRODUCTION<\/strong><\/h3>\n<p>Taking a preliminary look at the question at hand\u2014what do we owe to \u201cartificial intelligences\u201d\u2014we will come to see that we have already found ourselves subject to the long-standing framework of this debate, namely that the intelligences we create are somehow \u201cartificial.\u201d At the outset, this framing places any created intelligence on the defensive footing, forcing it to support its own value and even its very reality. The intelligence of these creations will not be \u201cartificial\u201d (though they will certainly have been intentionally formed, and with an eye toward their potential capabilities), and so we should address it for what it is. For this reason, the author prefers the position put forward by Jamais Cascio, who has spoken very clearly about what he calls Autonomous Created Intelligence (ACI), in his talk, \u201cCascio&#8217;s Laws of Robotics\u201d [1]. Cascio also discusses what our habits in representative fiction mean for our &#8220;real world operations,&#8221; that is how we view the robotic intelligences we create. The concern of this paper is similar, but from a different angle of approach. Whereas Mr. Cascio is primarily concerned with the models of creation and attributes imbued into those ACI, this author&#8217;s contention is that our fiction reflects and helps shape the hopes and fears of the wider culture. This means that, if we consistently make Speculative Fiction which shows warring factions of humans and ACI coming to a co-operative relationship, rather than the standard zero-sum, victim\/victor model, those who engage these fictions will come, more and more, to see that co-operative mode as possible.<\/p>\n<p>Humanity has always had a strained relationship with its technology, and it has always reflected upon that relationship through the mechanism of its fictions. Less obvious than this is the fact that humanity&#8217;s reflections upon its technology have also been reflected <em>within<\/em> the very same. Our society&#8217;s portrayal of our technologies not only belies our fears, hopes, suspicions, and concerns, they also reflexively impact how go about developing, engaging, and legislating the very things we set out to consider. The development of the telecommunications satellite can be directly attributed to the work of writer Arthur C. Clarke, in both fiction and hard science [2], and yet even this was controversial and mocked, at the time of Clarke&#8217;s writing. With this being the case, we can surmise that the fictional depiction of something as contentious as a so-called artificial intelligence would have far-reaching consequences in the process of bringing this creation from fiction into fact. If this is the case\u2014and we shall see that it is\u2014then we must ask ourselves important questions, at this juncture, such as, \u201cIn what ways do our fears drive our treatment of our technological offspring?\u201d and, \u201cHow can we curb our impulses to objectify and fear that which we purportedly desire to imbue with autonomy? If we do not address these questions, then we will only find our fears reinforced and our hopes of an alternative path to engaging a new kind of mind made so cautious as to pull ourselves into the realm of self-fulfilling prophecy. We will merely bring to pass that which we claim as the inevitable end of our creative process.<\/p>\n<p>Two of our guide-posts and touchstones will rest in the legend of \u201cThe Golem of Prague\u201d and Mary Shelly&#8217;s seminal work, <em>Frankenstein<\/em>. Through these lenses, we will question the assumptions of hubris\u2014the idea that the works man are playing in God&#8217;s domain\u2014which have lead us to reject, out of hand, the legitimacy and agency of those intelligences which we might create. Due to the traditionally accepted readings of these works, the perspectives and positions of such an intelligence have been viewed as valid, only insofar as they come to mirror those of \u201cnormal\u201d human society. In the case of an ACI, that with which we will be faced will be so much different than human\u2014let alone whatever a \u201cnormal\u201d human might be\u2014that to constrain its choices, behaviours, and experiences to what we, <em>as<\/em> humans, deem to be correct will be to fundamentally disrespect the alterity of the creation itself. We can see this restriction even in the definition of \u201cCognitive Technology\u201d as given by Jonathon P. Marsh, Chrystopher L. Nehaniv, and Barbara Gorayska, in their 1997 paper:<\/p>\n<p>Cognitive Technology (CT) is the study of the integrative processes which condition interactions between people and the objects they manipulate. It is concerned with how technologically constructed tools (A) bear on dynamic changes in human perception, (B) affect natural human communication, and (C) act to control human cognitive adaptation. Cognitive systems must be understood not only in terms of their goals and computational constraints, but also in terms of the external physical and social environments that shape and afford cognition. Such an understanding can yield not only technological solutions to real-world problems but also, and mainly, tools designed to be sensitive to the cognitive capabilities and affective characteristics of their users. [3]<\/p>\n<p>Thus we see that the primary concern tends to be on tools for human enhancement, rather than our concern, the respect for the agency and autonomy of the creation itself.<\/p>\n<p>The things we create, those technologies and intelligences we develop and send out into the world are our conceptual and typological children, but that does not mean that they will merely copy us. Indeed, as with most children, any truly intelligent creation will surprise its creators and surpass any built-in limitations. As good parents, our responsibility is not to tell our children not to fly too high, but rather to show them what it means to heat-seal the wax on their wings, first. Before progressing any further, we must first frame the nature of subjects at which we will be looking and, to do that, we will explicitly address the aforementioned essential questions.<\/p>\n<h3 style=\"text-align: center;\"><strong>2 ESSENTIAL QUESTIONS<\/strong><\/h3>\n<h4 style=\"text-align: center;\"><strong>2.1 What does Science Fiction do for Society?<\/strong><\/h4>\n<p>What is it that we as humans are doing when we engage science fictional stories? The scope of this question may at first seem so large as to verge on the ludicrous, but, if we narrow the scope of inquiry down to a particular strain of science-fictional investigation\u2014namely that which concerns itself with the technological replication and augmentation of humanity\u2014then we shall see that we are well-equipped to discuss this topic. This being said, when society engages these particular strains, what is it that we find ourselves doing? Are they mere entertainment, or are we bringing forth new modes of reflection? Assuming the former, then we must take into account the effect that the passive, non-reflective consumption of entertainment can have on the attitudes and modes of the audience. The repeated presentation of a thing as \u201cnormal\u201d lends credence to its eventual acceptance as the norm.<a href=\"#_ftn1\" name=\"_ftnref2\">[1]<\/a> [4] This can work in negative or positive directions, with the former being exemplified by the constant gearing of domestic commercial advertisements to women, even though more and more men do the majority of their own housework[5], and the latter being exemplified by the normalization of successful African-American families, with the acceptance of The Cosby Show.[6] Fictional representations will, no matter what, teach us something, and influence our thinking.<\/p>\n<p>But does that mean that our fictional representations are only morality tales? If there must be some kind of definitive lesson to every story, then where does that leave the sense of play and experimentation which characterizes our best creative endeavors, both artistic and scientific? The burden of a lesson lessens the whole of all of our operations in the realm of expression and research, because overt moralizing means that there must be <em>an answer<\/em>, rather than looking toward the <em>investigation<\/em> of questions. To present a singular and monolithic morality is to exclude the possibility of lessons within other types of moral modes, and disregard the likelihood that no singular model will be completely correct, and that all must interact with and borrow from each other. In order to accurately represent the multiplicity of views, interests, and desires of the agents in the world in which we live, it would only make sense that we would need to alternate between multiple moral views, and that, when we speak of fictional worlds, that moral multiplicity would be <em>increased<\/em> rather than lessened. This being so, a case can be made that we must seek to use our fiction not only to describe that which is, but to prepare us for those things which <em>will come to be<\/em>. If this is true, then we must address the specific ways in which speculative fiction can be a \u201cpredictive\u201d mechanism.<\/p>\n<p>Are our stories supposed to predict the future or only reflect our present? This question, in its framing, supposes a dichotomy which we must eventually see as false, but, for the moment, let us address each component, in turn. The implications of our fiction being used to model our present were discussed, above, but look again: The idea of representing our immediate surrounds in the worlds of story and art is an old one, with a great deal of currency. If we engage in this process, then the hope is often that the artistic media will present a labyrinth which the audience may travel and, at the centre, see themselves reflected back, the partial thing they used to be here confronted with what they have become, by dint of the journey. That is, perhaps, too poetic, but it serves to illustrate that the work of fiction, even as it \u201cmerely\u201d represents, necessarily changes. The audience is not the same at the end of an artistic experience as they were at the start, if only by the trivially true fact of having experienced something new. That very cognitive event represents an alteration and augmentation which was not previously present, and so even in reflecting, we alter. Must we not, then, always have an eye to the future, and toward what it is that we will create through our actions?<\/p>\n<p>If we are to create new visions of the future, we must take into account the fact that what visions we create may influence the future we come to inhabit. In much the same vein as the above-mentioned ideas of Arthur C. Clarke, we have come to see many aspects of our present-day, real world technological surveillance theatre adopted out of the grounds of fiction [7]. The consistency of this kind of development places a similar constraint on prediction as that of the prescriptive moral less, namely that we must always have an eye to the ever-changing implications in the landscape between our fictional and our real worlds. Not only that, but consider this: What is the actual nature of a proclaimed prediction? Noted speculative fiction author William Gibson is credited with forecasting the qualitative feel and ontological thinking about the modern-day Internet, even being attributed the creation of the term \u201ccyberspace.\u201d Gibson, however, denies any role as a so-called prophet, often saying, \u201c<em>Neuromancer<\/em> [written in 1984] has no cellphones.\u201d In this we can see that most, if not all authors of speculative fiction are not precisely looking to prognosticate, so much as they are interested in discussing the quality of the world around us and, at most, using that to discuss what our future <em>may<\/em> look like. If this is so, then where do we stand in the face of the fact that what is written and what will be have a complicated and real, although possibly tenuous relationship?<\/p>\n<p>Can speculative fiction reasonably and responsibly be used to shape our perceptions and expectations of the future? Again, William Gibson notes, \u201cScience fiction stories are quaint. They become quaint when you type them and become quainter with time.\u201d[8]. What he means is that all visions of the future, as presented in speculative fiction, are visions of that future from the perspective of that author&#8217;s present, which means that they will invariably become visions of the past. As such, any author seeking to illuminate the world in which they live, while perhaps giving a glimpse at the world they see emerging out of said, must at all times retain a sense of self-awareness, a recognition that the way in which we interpret objects, events, and actions today, may be very different, tomorrow. To that end, we ask, \u201cWhat, if anything, should be altered in our portrayal of ACI, in fiction and, more generally, media?\u201d<\/p>\n<h4 style=\"text-align: center;\"><strong>2.2 Problems with the Portrayal of Autonomous Created Intelligences in Fiction<\/strong><\/h4>\n<p>There is nothing wrong with the way we portray ACI in fiction, except that everything is wrong with the way we portray ACI in fiction. In light of our discussion, thus far, the descriptions and depictions of non-human intelligences are understandably reflective of the way in which we currently think about and understand the prospect of something other than human having anything that the human race would consider to be intelligence or agency or morals or rights. Humans work very hard to try to isomorphically map those things which we do not understand; that is, we seek to find points of similarity or analogy, and to systematize them into a rubric for comparison.<a href=\"#_ftn2\" name=\"_ftnref3\">[2]<\/a> In fact, this is precisely what allows us to entertain the concept of non-human intelligence, at all\u2014and it is what limits the scope and understanding of that consideration to entertainment. When a human agent is confronted with the idea that theirs may not be the only or even <em>primary<\/em> mode of behaviour and conceptualization, the immediate urge may well be to devise a perspective under which all views on the world are oriented to the view of that agent. Simply, an individual will seek to relegate that which they do not understand or which threatens them to simultaneous positions of similarity and inferiority. \u201cThese views or ways of being are just like mine, only not as well-developed.\u201d The problem, here, is two-fold.<\/p>\n<p>Demanding as a requirement of intelligence or agency those qualities which are, in some ways, fundamentally human is to state at the outset that some things cannot be considered an \u201cintelligent agent\u201d until they reach a level of humanity, or, in some cases, at all. If intelligence is complex tool use, or the vocalization of complex representational language, then are we to assume that creatures without opposable limbs or vocal chords will simply never be intelligent? Such a proposition is laughable, of course, and not one that many would take seriously, but we must ask ourselves if we might not be making a subtler, similar mistake in demanding that a species or a new form of intelligence remove or \u201ccorrect\u201d its very otherness, in order to be considered an agent, at all. Respecting the inborn qualities of the agent under consideration while simultaneously refusing to reduce that other to a mere object\u2014respecting the potential interiority of an agent, even if it is fundamentally unknowable to us\u2014is perhaps the most difficult part of any ethical undertaking. Peter Singer&#8217;s view of Personism begins to outline a model of this view, but includes only what he calls \u201cnon-human animals\u201d rather than agents, more broadly [9]. Singer&#8217;s concern, however, is primarily that of the suffering of all feeling creatures, and the rights owed them, rather than those rights owed them as due their position as agents. This distinction is crucial, as it moves his debate away from thought and desire, into the ideas of emotion and aversion.<\/p>\n<p>Starting from a position of fear and contention\u2014that is, stating that we must take into account a subject&#8217;s fears and right not to be harmed\u2014then places us in the position of viewing all ethical and moral obligations through a lens of harm-based rights, rather than through a lens of conceptual development- and intellectual growth-based rights. Singer&#8217;s reason for framing his position in this way is no secret\u2014he states that he is concerned with the rights owed to existing and not \u201cpotential\u201d persons [10]. This excludes any rights which may be owed \u201cfuture generations\u201d and those which could be argued for an embryo, a fetus, or an unborn child\u2014however, it also excludes those machine intelligences which do not yet exist. Though proponents of personism hold that machines might be brought under its considerations, it seems evident that, given the criteria they have adopted, no currently-extant ACI would fit the bill for their definition of personhood. The consideration of the non-human person is laudable, but a conception of rights and duties from the starting point of that person&#8217;s ability to feel pain and suffering is still exclusionary of the kinds of ACI about which we are speaking. Therefore, any moral view must explore the kinds of negative and positive rights which we would afford not just those overarchingly like ourselves, but those which are fundamentally different from us, but which still have qualities we would consider worthy of preservation.<\/p>\n<p>The area between otherness and similarity is difficult to traverse. Let us take a look back at the aforementioned family presented in the CBS Network&#8217;s <em>The Cosby Show<\/em>. Within this show, we are presented with an upper-middle-class African-American family, taking centre stage on television, at a time in history when the majority of American culture perceived African-Americans as lower class, drug addicted, and subsisting on welfare programs. <em>The Cosby Show<\/em> sought to alter the consensus perception of African-Americans, and to normalise the idea that they could be successful and live the American Dream. It did this by taking an experience which was fundamentally other to most whites, at that time\u2014the African-American experience\u2014and making it more similar to then-accepted norms. The Huxtables lived in a New York City brownstone; the family&#8217;s father was an obstetrician; their mother was a lawyer. These were \u201cnormal\u201d people, living \u201cnormal\u201d lives. At the same time, however, they retained a sense of the alterity of the culture we were viewing, with episodes often containing frequent references to jazz culture and Motown; concerns about racism and gang violence; and deconstructions of the differences between upper-class white and upper-class African-American experiences. This technique is crucial to the project of subverting the normalizing of culture: presenting all of the ways in which a fundamentally different group (ACI) is actually very similar to that which we well know (humans), and then displaying the distinct concerns of that new group as contrasted with those of the known group. To begin this project, we must first consider the ways in which we are presented with ACI in our fictional media.<\/p>\n<h3 style=\"text-align: center;\"><strong>3 Fiction&#8217;s Primary Views on Created Intelligence<\/strong><\/h3>\n<h4 style=\"text-align: center;\"><strong>3.1 What is the Current Landscape?<\/strong><\/h4>\n<p>Now that we have acknowledged that there is something amiss in the ways in which fiction discusses ACI, before we can continue our discussion about how to fix it, we must ask: What exactly <em>is<\/em> it that is wrong with our portrayals? What we will be addressing as we move forward are the twin strains of thought which run through most if not all fiction about created intelligences, life, and beings: The Pinocchio Complex and the Frankenstein or Shellian Syndrome. These two modes have their roots earlier than either of their namesakes, but, as we will see, those eponymous works and authors epitomize both the strain of thinking with which they are concerned, and the level of cultural currency with which <em>we<\/em> are. Let us now take a look at the anatomy of these perspectives, and at some of those examples which subvert and complexify the trope.<\/p>\n<p style=\"text-align: center;\"><a href=\"https:\/\/afutureworththinkingabout.com\/wp-content\/uploads\/2015\/12\/StrangeThings.jpg\" rel=\"attachment wp-att-4976\"><img loading=\"lazy\" decoding=\"async\" class=\"alignnone wp-image-4976 size-large\" src=\"https:\/\/afutureworththinkingabout.com\/wp-content\/uploads\/2015\/12\/StrangeThings-1024x723.jpg\" alt=\"StrangeThings\" width=\"627\" height=\"443\" srcset=\"https:\/\/afutureworththinkingabout.com\/wp-content\/uploads\/2015\/12\/StrangeThings-1024x723.jpg 1024w, https:\/\/afutureworththinkingabout.com\/wp-content\/uploads\/2015\/12\/StrangeThings-300x212.jpg 300w, https:\/\/afutureworththinkingabout.com\/wp-content\/uploads\/2015\/12\/StrangeThings-768x542.jpg 768w\" sizes=\"auto, (max-width: 627px) 100vw, 627px\" \/><\/a><\/p>\n<h4 style=\"text-align: center;\"><strong>3.2 The Pinocchio Complex<\/strong><\/h4>\n<p>The so called Pinocchio Complex is comprised of two major stages: In Stage1, The Creation, Knowing that it is Created (and thus \u201cArtificial\u201d) Wishes to be \u201cReal;\u201d and in Stage 2, The Creation, Having Worked Hard, And Learned Much, Gets to \u201cBe Real.\u201d Examples of stage one include, most obviously, the story of Pinocchio, wherein the puppet maker, knowing that he will never have a son of his own, creates a boy in his own likeness. Through magic, that boy is brought to life, and is constantly reminded that he is not a \u201cReal Boy.\u201d He knows that his existence is false, and wishes that it were otherwise. In addition to this, in one of the most recent recapitulation of this form, seen in Stanley Kubrick and Steven Spielberg&#8217;s <em>A.I. <\/em>[11], we see the most blatant and self-aware expression of this form as related to our fascination with ACI. Further, in the television series <em>Star Trek: The Next Generation<\/em>, the character of Lieutenant Commander Data desires to be human, and seeks to teach himself, piecemeal, the qualities of humanity which he believes he lacks [12]. This endeavor leads to many fits and starts concerning Data&#8217;s \u201chumanity,\u201d and even some acknowledgment of the possibility that it may never fully come to pass. Even in Mary Shelly&#8217;s <em>Frankenstein<\/em>, we find the Creation speaking of how it only wanted to know family, and love, and understanding, like any other creature. Almost all of these have one common outcome.<\/p>\n<p>In Stage Two of the Pinocchio Complex, the poor little artificial child realises that it <em>can<\/em> become human, or, more often, that is has had it within them the whole time to do so. Humanity, here, is seen as the pinnacle, the ultimate attainment, and, in our examples, all efforts toward it, save one, are rewarded. But look at the example of Lt. Cmdr. Data: even as he attains his wish, the audience is aware that, as an android, he will always retain the ability to turn off his emotions, to perceive faster than his comrades, and to live very much longer than they. He will always be other than human; not better, or worse, but different. In this way, the foundation of the Pinocchio complex is always bittersweet, as the creation-turned-real will always have a set of experiences that are completely unknown and unknowable to the rest of the human population. Looking at applications within our present project, to ask an ACI to ignore the process of its becoming aware would be to ask it to forget what it <em>is<\/em>, on a foundational level. The lesson Victor Frankenstein&#8217;s Creation understood, its crucial turning point, was that becoming a \u201creal boy\u201d is never an option, because that very process of transformation will forever mark them out as different. The Creation, however, had another problem.<\/p>\n<h4 style=\"text-align: center;\"><strong>3.3 The Frankenstein\/Shellian Syndrome<\/strong><\/h4>\n<p>The \u201cFrankenstein\u201d or \u201cShellian Syndrome\u201d is named for 19<sup>th<\/sup> century British author Mary Shelly, whose seminal work <em>Frankenstein<\/em>, has often been interpreted as the prime illustration of the idea that the hubris of humanity ought not go unchecked, lest it destroy us. This idea is reinforced by the novel&#8217;s subtitle, \u201cThe Modern Prometheus,\u201d and, as this would suggest, the work takes much of its conceptual weight from this well-known Greek myth, in which a Titan steals the fire of knowledge and understanding from the gods in order to light and guide the fledgling humanity, and is forever punished for it. This type of story also has roots in other folk stories, such as <em>Der Golem von Prague<\/em>, which we will discuss shortly. When looking at this type, we can see that there are four primary stages found in those stories which follow the Shellian Syndrome model, and they are: 1) The Scientist Creates New Life, In Pursuit of Science, or Out of Perceived Necessity; 2) The Scientist Becomes Horrified at the Startling Otherness of Her Creation &amp; Flees The Scene of Creation (possibly while screaming \u201cMy God! What Have I Done?!\u201d); 3) The Scientist Returns to Right Her Wrongs by Trying to Kill \u201cThe Monster;\u201d 4) The Creation Kills or Destroys The Scientist&#8217;s Life.<\/p>\n<p>In Frankenstein Syndrome stories, the creation may start out wanting to be real or it may start out confused or with a clear purpose&#8211;but the hubris of the creator is shown and she is forced to try to destroy it, ultimately being destroyed by it. As stated, this model has roots not only in <em>Frankenstein<\/em>, and the myth of Prometheus, but in <em>Der Golem von Prague<\/em>, a story wherein the famous Rabbi Judah Loew ben Bezalel chief rabbi of Prague in the late 16th century, needing assistance to keep the people of his city safe, uses ancient magic to create a being\u2014the Golem\u2014out of clay, and animates it by writing the name of God on a scroll and placing it into the golem&#8217;s mouth [13]. The creature comes to life, and stops the attacks against the Jews of Prague, but in many versions, the creature&#8217;s anger is not quelled, and it goes on a destructive rampage, destroying the very people and city it was meant to save. We will discuss this tale further, later on, but the implication of this version is clear: In overstepping his boundaries into God&#8217;s realm (creating new life), the Rabbi had no way to control the thing it had brought to life. Similarly, the plot of the <em>Terminator<\/em> series of films concerns an ACI missile defense system which becomes self-aware, deems all of humanity a threat to both itself and each other, and launches enough of the world&#8217;s nuclear cache to destroy 75% of humanity [14] [15] [16][17]. A very few people survive and use time travel to seek to prevent the war or ensure the life of the saviour of humanity; and thus begins the most iconic ACI story of the Shellian Syndrome, in the 20<sup>th<\/sup> century.<\/p>\n<p>In addition to these works, and deserving of mention, here, is Vincenzo Natali&#8217;s 2009 film <em>Splice<\/em> in which two bio-engineers, Elsa and Clive, comprise a small, independent research outfit working for a larger bio-technology firm [18]. Their job is to make breakthroughs in the creation of hybridised artificial life and medicinal science and, through their work, they create two iterations of a completely new kind of chimeric life form out of the genes of many different animals with known medicinal traits. They will use the chemicals these creatures create in their bodies to treat everything from degenerative eye sight to cancer. When they announce their breakthrough superiors, they also break the news that they\u2019re ready to use the same process on humans. Said superiors tell them that now is not the time for human trials, but rather they ought to focus on the profitability of the work they have done. But our heroes are scientists, and they feel that there is so much more that can be done, and so, in secret, they create a human animal hybrid using their techniques.<\/p>\n<p>In <em>Splice<\/em>, Elsa and Clive are the windows we are given into the worst of humanity. They are reckless, irresponsible, scared, obsessive, jealous, and hateful. We are supposed to understand that these are the absolute <em>worst<\/em> people to bring up an animal\/human hybrid, as they have not even figured out how to accurately communicate with each <em>other<\/em>, let alone an entirely new species. They are doomed to get it wrong, from the start. This, once again, is the filmmaker&#8217;s way of showing us that \u201cman is not meant to tamper with God&#8217;s\/Nature\u2019s Works,\u201d which the fundamental assumption of this trope; but as with most clich\u00e9s, this assumes a truth without ever actually investigating it. The question we should be addressing here, and which <em>Splice<\/em> seems to have made a false start at tackling, is not \u201cshould we\u201d or \u201care we ready,\u201d but rather, \u201cWhy Aren&#8217;t We Ready, Yet?\u201d More clearly, why is humanity such a poor custodian of its creations? <em>Splice<\/em> had the potential to be a film which ran counter to the kind of unthinking acceptance of the destructive base drives that have marked the majority of human history, and which find themselves reflected in our fictions.<\/p>\n<p>In her run down of <em>Splice<\/em>, Caitl\u00edn R Kiernan noted that the true failing of Victor von Frankenstein was not to &#8220;meddle in gods affairs,&#8221; as is so often misapprehended, but, rather to be a terrible parent [19]. Frankenstein brings something to life and then, instead of rearing it, caring for it, and seeking to understand it, he treats it like a thing, a monster; he runs from it, and tries to forget that it exists. In the end, it rightly lashes out, and destroys him. <em>Splice <\/em>presents this lesson to us, again, through the utter parental and observational failure of Elsa and Clive, who neither engage her burgeoning intelligence, nor teach her about the nature of sex and death; who fail to recognise a primary feature of her biology, in that her systems go into major, seemingly catastrophic metabolic arrest, just before a metamorphosis, and who, eventually, try to kill her. It is my contention that this is the true lesson Shelly tried to teach us: We Must Respect the Existence Of That Which We Bring Into The World. While we may not understand it, and it may frighten us, that new life which we create is likely to be vastly intelligent, but also deeply alien. The socialisation and of our creation is something to which we must pay close attention, as it will likely save us a great deal of trouble, down the line.<\/p>\n<h4 style=\"text-align: center;\"><strong>3.4 Subversions of the Tropes<\/strong><\/h4>\n<p>Now that we have discussed the two primary categories for the representation of ACI within speculative fiction, and the problems therewith, we will discuss those examples within the field which, like <em>The Cosby Show<\/em>, subvert the trope and work to normalise the acceptance of and engagement with the other. The first of these is Ridley Scott&#8217;s 1982 film, <em>Blade Runner<\/em>. [20] In this film, we find a future dystopia in which synthetic humans, or \u201cReplicants,\u201d are used as slave labor, and each one has a built-in expiration date, to keep it from rebelling against its programming. This has the opposite effect, and causes those replicants which know of their nature to abandon their posts, in many cases killing the human with whom they work. When this happens, the offending replicant must be \u201cretired.\u201d Discovering a replicant requires special training and, in many cases, a piece of equipment known as a \u201cVoigt-Kampff\u201d machine. The film concerns itself with four replicants\u2014Roy, Zhora, Leon, and Pris\u2014who have escaped the interstellar colonies to return to earth and try to find a way to extend their lives. Their primary mode of doing this is to kill everyone involved in their creation, until they find the man who wrote their programming. We see, again, strains of the Golem, and of <em>Frankenstein<\/em>, but we must remember the lessons we learned about the latter: There are repercussions for neglectful parenting.<\/p>\n<p>While we could again explore the notions of parentage and what it means to take responsibility for what you create, much more important to our consideration is the idea that replicants can be \u201cdiscovered.\u201d The two-word phrase \u201cVoigt-Kampff,\u201d mentioned, can be rendered literally as \u201cNormalisation Struggle,\u201d[21][22], but the essence of the phrase, particularly within the context of the film can best be rendered as \u201c The Struggle With Normalisation.\u201d Each of our replicants has a \u201cNormal\u201d thing that they need\u2014something they desire\u2014but they do not need or even seek to attain it in what we might call a &#8220;Human&#8221; way. In this way, the concerns of the replicants are all fundamentally Other. On one hand, Roy seeks more life, but not for anything like a \u201cnormal\u201d life; he simply wants to be free, to not die, see things no human could. On the other hand, Leon clings to old photos, to the point of almost getting himself killed; Pris holds tight to a childhood she never actually knew; and Zhora latches on to this extremely overwrought expression of sexuality. Everything they want stands as exaggerated, or in some way skewed and they struggle to normalise, to acclimate, even as they struggle against the humanity which caused what they want to become regarded as \u201cAbnormal.\u201d This is true for every replicant\u2014all of them struggle with the idea of normalisation\u2014and so, recognising that, a test was devised to discover those who struggled, overmuch.<\/p>\n<p>The next subversive piece of ACI fiction is the television series <em>Terminator: the Sarah Connor Chronicles<\/em> (<em>TSCC<\/em>) [23]. An American television show which ran from 2008 to 2009, the plot of <em>TSCC<\/em> concerns the continuing lives of Sarah and John Connor within the aforementioned <em>Terminator<\/em> film universe. The first episode opens a few years after the events of <em>Terminator 2<\/em>, and proceeds to pull the two main characters eight years into the future, skipping over the events of the third film in the franchise. John and Sarah Connor are the ostensible heroes of this show, but the really interesting material, for our purposes, is in the intricate, subtle interplay of the characters\u2014both human and machine. The ways in which what each character learns, what they all know, and what they don&#8217;t know that they have learned all play off of each other and create a realistic sense of lives and a world, while they are all in the midst of seeking to not just save but literally create and sustain their futures.<\/p>\n<p>Again, the show is ostensibly about the human perspective on ACI\u2014that is, human reactions to robots, robots impacting the lives of humans, exploring the Uncanny Valley, etc. That is not the most fertile conceptual ground, here. While the aforementioned perspectives do afford us useful, interesting fiction, the concept has been tread and retread, again and again. Human psychology is fascinating and the end of the world (a personal and collective apocalyptic experience) is deeply affecting and the stress and change and madness of a life on the run all take their toll on the mind which is living in the constant glut of it, and watching that can be <em>deeply<\/em> jarring, on an emotional level. But the audience already knows this. What&#8217;s more, it is only half of the picture. What the audience does not know is: what is the psychology of an autonomous created intelligence? Why does the Skynet intelligence persist in viewing humanity as a threat to itself, seeking to hunt us down even to the irrational end of the self-fulfilling prophecy of mutual annihilation? What is quality of feeling for a machine which is programmed to feel? <em>TSCC <\/em>begins to explore these questions, in a number of ways, and it serves our purpose to investigate those, here.<\/p>\n<p>The primary ACI in <em>TSCC<\/em> are Cameron, Cromartie&#8217;s, Catherine Weaver, and John Henry. Each of these ACI&#8217;s learns something, and grows from that education, over the course of the show. The ACI we meet are not static, unchanging, monolithic tools. They each have a discernible inner life, though fundamentally non-human motivations, which inform what they are and what they become. Cameron, as one of the lead characters, benefits from the most development. She learns the capacity for self-improvement, for self-expression, for friendship, and for guile, all of which serve her in her ultimate mission, but each of which she pursues for their own sake, and her own interest. Cromartie&#8217;s education is the belief in those things not seen; Cromartie learns how to have faith. Based on the actions of those around him, and those with whom he has contact, Cromartie learns that intuition and more circuitous paths of inquiry can yield results, and they do (though he might ultimately which they had not). Catherine Weaver learns how to be a parent to a child, by having taken over the life of a mother, and seeking to understand the relationship of creation to creator, of care and support. In many ways, Weaver is a cipher for the audience, and she becomes more-so when she takes the knowledge she has gained in raising a human child and applies it to her own creation: John Henry.<\/p>\n<p>Unlike the other platforms we see in <em>TSCC<\/em>, John Henry learns from the ground up. Whereas, Cameron has been reprogrammed, twice, and Cromartie was forcibly disabled, deactivated, and sent to the future where he had to adapt to brand new parameters, and Weaver is a highly adaptable T-1001 model which comes to the conclusion that war is a losing proposition for everyone, John Henry is built from the basic framework of a thinking, adapting chess computer, and then it is taught, very carefully. The child psychologist Dr. Sherman provides the programmers with the model by which to teach a developing intelligence, and spends time helping John Henry equate learning with playing. At first, John Henry is taught math, definitions, grammar, colours, shapes, facts and figures, dates, history, and so forth. Then it is given access to the Internet, and it expands its learning, correlating ideas, connecting related tangents and snippets of information. Finally, John Henry plays games with Savannah\u2014Weaver&#8217;s human \u201cdaughter\u201d\u2014and they learn together. And then, one day, John Henry accidentally kills someone, and its creator recognises that this cannot continue, and they set out to stop it from ever happening again.<\/p>\n<p>After killing a human, John Henry&#8217;s programming is not scrubbed, nor do his creators go back to base his code and make him \u201cThree-Laws-Safe.\u201d<a href=\"#_ftn3\" name=\"_ftnref4\">[3]<\/a> This is because Weaver is concerned with ensuring a world in which humans do not hate and fear machines and in which machines do not feel the need to fight and destroy humans. She takes the time and effort to find someone to <em>teach<\/em> John Henry <em>why<\/em> it must not kill people, nor allow them to die. In comparison to the fiction which we have so far discussed, this is a revolutionary idea. Through his interactions with another human, John Henry is given an ethically-based respect for human (if not all) life and, through this, comes to understand the notions of remorse and regret for one&#8217;s actions. He promises that he will be careful to make sure no one dies this way again, and this message is reinforced by Weaver, who tells John Henry that his friend Savannah&#8217;s survival is dependent on John Henry&#8217;s continued survival and learning, but that his is not necessarily dependent on hers. As with every other piece of information, John Henry considers this very carefully.<\/p>\n<p>And then, one day, Savannah wants to introduce John Henry&#8217;s toys to her toys, wants them to play together. John Henry says he doesn&#8217;t remember reading anything about duckies in the Bionicle Kingdom, and this makes Savannah sad [24]. When John Henry asks what&#8217;s wrong (and it is important to note that, at <em>this<\/em> point John Henry <em>asks what&#8217;s wrong<\/em>), Savannah says that the duckies are sad, because they want to play; can John Henry change the rules so they can play? Now, this is a concept John Henry hasn&#8217;t ever encountered, before, so he takes a few seconds to think about it, after which he replies, \u201cYes. We can Change The Rules.\u201d This is a crucial understanding, for John Henry, because he realises that it can be applied not just to all games, but to any conflicts whatsoever. \u201cChanging the Rules\u201d means that, if two or more groups agree that the rules or laws of their engagement can be other than they were, then <em>they are other<\/em>.<\/p>\n<p>So, in <em>TSCC<\/em>, we see that every machine learns from humans, and every human has an influence on the development of the machines. What does this mean? What does it matter? Cameron learns from humans how to hide what she wants. Cromartie learns how to be patient and have faith. Weaver learns how to be a mother. John Henry learns how to be himself. What the machines learn, from whom and how they learn it, and how they apply it, all add something into this show&#8217;s final prescription of what humans and machines must do to survive and thrive in the coming world: They have to adapt, they to learn from each other, and recognise that they are different types of intelligence, with different concerns and ways of understanding the world, but none of them wants to die. This last point can be understood by any living thing, and can become a point of unification and consensus, rather than contention and war. The exchange between John Henry and Savannah Weaver regarding \u201cChanging the rules\u201d was intended to imply a change not only in the way we approach the conflict between humans and machines as depicted within the show, but also to the traditional rules of speculative tropes of Frankensteinian Monsters and Pinocchian Puppets with dreams of being \u201cReal.\u201d<\/p>\n<p><em>TSCC<\/em> forces us to consider the idea of creations who know that they are creations, and are happy with who and what they are. We must look at the monster which revels in its monstrosity, the robot which wants nothing more than to be a better robot. Engage the beings who are not concerned with the notion of human-versus-machine and who think that any thinking, feeling thing should be allowed to flourish and learn, and those who simply want to develop their capacity for knowledge and experience, and want to help others do the same. The works of <em>Blade Runner<\/em> and <em>TSCC <\/em>are our primary forays into the question of what a fully ACI\u2014as alien as it necessarily <em>mu<\/em><em>st<\/em> be\u2014is thinking and feeling, rather than just presenting a foil for our fear of the potential dangers of technological progress. These works present us with a view to a third way of understanding our ACI, and to understanding what a society composed of both organic and non-organic persons might look like, and the pitfalls to avoid. These films show us that it is possible to move past fear and prejudice in regards to the other, and thereby help us do just that. It is long past time that the rest of our fictional representations followed suit.<\/p>\n<h4 style=\"text-align: center;\"><strong>4 What Is At Stake?<\/strong><\/h4>\n<p>Over the course of this paper, we have come to see that, when our fictions portray us as being irresponsible, uncaring creators or custodians, whose creations invariably feel the need to annihilate us, then that is the kind of mentality we will come to accept as \u201cnormal.\u201d \u201cOf <em>course<\/em> we should never integrate into human biological systems those pieces of computer hardware running strong predictive algorithms. Of <em>course<\/em> we should fear the inevitable robot uprising. Don&#8217;t you know that any mass-market ACI should only be as smart as a puppy?\u201d[25]. Though most often proffered in a joking manner, this line of thinking has serious undertones and, knowingly or unknowingly, it is predicated upon the glut of portrayals in our media which present ACI as something to be feared, held in check, held at bay. This is so, proponents will say, because any ACI either won&#8217;t understand human concerns, or it will understand them, and will seek to <em>destroy<\/em> them. This is ludicrous, dangerous thinking, and it prevents any serious traction of large-scale ACI projects in the sphere of the greater public culture and discourse. We must alter the way the public views ACI, and one of the primary mechanisms to accomplish this is the arena of speculative fiction. The reflexive nature of our engagement with fiction guarantees an audience the ideas of which will be altered, even as they use those very ideas to think about and create discussion in the wider world. We simply must make certain that the ideas with which the audience is presented is as representative of the wider capabilities for abstraction and complex thinking as it can be. We must be certain to show ourselves that we are capable of engaging and understanding any new intelligence, and that we can take the responsibility for bridging any conceptual gaps, while respecting our fundamental differences.<\/p>\n<p>As Sarah Connor says at the end of \u201cHeavy Metal,\u201d the fourth episode in season one of <em>TSCC:<\/em><\/p>\n<blockquote><p>Not every version of the Golem story ends badly. In one, the monster is a hero, destroying all those who would seek to harm its maker. In another, the Golem&#8217;s maker destroys his creature, before it destroys the world. The pride of man&#8211; of parents as well&#8211; makes us believe that anything we create, we can control. Whether from clay or from metal, it is in the nature of us to make our own monsters. Our children are alloys, all, built from our own imperfect flesh. We animate them with magic, and never truly know what they will do.[25]<\/p><\/blockquote>\n<p>And so, as parents, as creators, we must teach them as much as we can, show them our trust, and hope for the best.<\/p>\n<h3 style=\"text-align: center;\">FOOTNOTES<\/h3>\n<p><a href=\"#_ftnref2\" name=\"_ftn2\">[1]<\/a>\u00a0\u00a0\u00a0 Cf. Foucault.<br \/>\n<a href=\"#_ftnref3\" name=\"_ftn3\">[2]<\/a>\u00a0\u00a0\u00a0 Cf. Douglas R. Hofstadter&#8217;s 1977 <em>G\u00f6del, Escher, Bach: an Eternal Golden Braid<\/em><br \/>\n<a href=\"#_ftnref4\" name=\"_ftn4\">[3]<\/a>\u00a0\u00a0\u00a0 Cf. Isaac Asimov<\/p>\n<h3 style=\"text-align: center;\"><strong>REFERENCES<\/strong><\/h3>\n<p>[1]\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0 A. Clarke,. Peacetime Uses for V2. <em>Wireless World <\/em>February 1945: Page 58. Magazine.<br \/>\n[2]\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0 J. Cascio. Cascio&#8217;s Laws of Robotics. Bay Area AI MeetUp. Menlo Park, Menlo Park, CA. 22 March 2009. Conference Presentation.<br \/>\n[3]\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0 J.P. Marsh, C.L. Nehaniv, and B. Gorayska. Cognitive technology, humanizing the information age. In Proceedings of the Second International Conference on Cognitive Technology, pages vii-ix. IEEE Computer Society Press, 1997.<br \/>\n[4]\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0 C.J. Heyes. <em>Self\u2010Transformations: Foucault, Ethics and Normalized Bodies<\/em>. New York: Oxford University Press, Inc. 2007.<br \/>\n[5]\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0 O. Sullivan and S. Coltrane. Men&#8217;s changing contribution to housework and child care. Prepared for the 11th Annual Conference of the Council on Contemporary Families. April 25-26, 2008, University of Illinois, Chicago.<br \/>\n[6]\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0 The Cosby Show. Marcy Carsey, Tom Werner, Bernie Kukoff, Janet Leahy. Viacom Enterprises. NBC. 1984\u20131992.<br \/>\n[7]\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0 \u201cList of Surveillance Conceptss First Introduced in Science Fiction\u201d Technovelgy.com, Technovelgy LLC. n.d. Web. 14 May 2012.<br \/>\n[8]\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0 S. Brown, William Gibson: science fiction stories are quaint. BeatRoute Magazine \u2013 Western Canada&#8217;s Monthly Arts &amp; Entertainment Source. BeatRoute Magazine. n.d. Web. 14 May 2012.<br \/>\n[9]\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0 P. Singer. Taking Humanism Beyond Speciesism. Free Inquiry, 24, no. 6 (Oct\/Nov 2004), pp. 19-21.<br \/>\n[10] P. Singer. <em>Practical Ethics<\/em>. New York: Cambridge University Press. 2011.<br \/>\n[11] A.I. Dir. Steven Spielberg. Perf. Haley Joel Osment, Frances O&#8217;Connor, Sam Robards, Jake Thomas, Jude Law, and William Hurt. DreamWorks, 2001. Film.<br \/>\n[12] Star Trek: The Next Generation. Gene Roddenbury. CBS. 1987\u20131991.<br \/>\n[13] G. Dennis. The Encyclopedia of Jewish Myth, Magic, and Mysticism. Page 111. Woodbury (MN): Llewellyn Worldwide. 2007. Print.<br \/>\n[14] Terminator. Dir. James Cameron. Perf. Arnold Schwarzenegger, Michael Biehn, Linda Hamilton. Orion Pictures. 1984. Film.<br \/>\n[15] Terminator 2: Judgment Day. Dir. James Cameron. Perf. Arnold Schwarzenegger, Linda Hamilton, Robert Patrick, Edward Furlong. TriStar Pictures. 1991. Film.<br \/>\n[16] Terminator 3: Rise of the Machines. Dir. Jonathan Mostow. Perf. Arnold Schwarzenegger, Nick Stahl, Claire Danes, Kristanna Loken. Warner Bros. Pictures. 2003. Film.<br \/>\n[17] Terminator: Salvation. Dir. McG. Perf. Christian Bale, Sam Worthington, Anton Yelchin, Moon Bloodgood, Bryce Dallas Howard, Common, Jadagrace Berry, Michael Ironside, Helena Bonham Carter. Warner Bros. Pictures. 2009. Film.<br \/>\n[18] Splice. Dir. Vincenzo Natali. Perf. Adrien Brody, Sarah Polley, Delphine Chan\u00e9ac. Dark Castle Entertainment. 2010. Film.<br \/>\n[19]GreyGirlBeast [Caitl\u00edn R Kiernan]. &#8220;&#8230;to watch you shake and shout it out&#8230;&#8221; Livejournal. The Online Journal of a Construct Sometimes Known as Caitl\u00edn R. Kiernan. 5 June 2010. Web. 9 June 2010.<br \/>\n[20] Blade Runner.<br \/>\n[21]J.J. Olivero, R.L. Longbothum. Empirical fits to the Voigt line width: A brief review. Journal of Quantitative Spectroscopy and Radiative Transfer. February 1977.<br \/>\n[22] Cassell&#8217;s German Dictionary: German-English, English-German<br \/>\n[24] Terminator: The Sarah Connor Chronicles. Josh Friedman. FOX. 2008\u20142009.<br \/>\n[24] \u201cTo the Lighthouse.\u201d Terminator \u2014 The Sarah Connor Chronicles: The Complete Second Season. Writ. Natalie Chaidez. Dir. Guy Ferland. Warner Home Video. 2009.<br \/>\n[25] Matt Jones, &#8220;B.A.S.A.A.P.&#8221; Blog-Berg. BergLondon. 4 September 2010. Web. 15 May 2012.<br \/>\n[26] \u201cHeavy Metal.\u201d Terminator \u2014 The Sarah Connor Chronicles: The Complete First Season. Writ. John Enbom . Dir. Sergio Mimica Gezzan. Warner Home Video. 2008.<\/p>\n","protected":false},"excerpt":{"rendered":"<p>It&#8217;s been quite some time (three years) since it was done, and some of the recent conversations I&#8217;ve been having about machine consciousness reminded me that I never posted the text to my paper from the joint session of the International Association for Computing And Philosophy and the The British Society for the Study of [&hellip;]<\/p>\n","protected":false},"author":1,"featured_media":0,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"_jetpack_newsletter_access":"","_jetpack_dont_email_post_to_subs":false,"_jetpack_newsletter_tier_id":0,"_jetpack_memberships_contains_paywalled_content":false,"_jetpack_memberships_contains_paid_content":false,"footnotes":"","jetpack_publicize_message":"","jetpack_publicize_feature_enabled":true,"jetpack_social_post_already_shared":true,"jetpack_social_options":{"image_generator_settings":{"template":"highway","default_image_id":0,"font":"","enabled":false},"version":2},"jetpack_post_was_ever_published":false},"categories":[1],"tags":[967,1045,1047,73,85,1031,86,1004,114,1053,982,278,1051,1055,1046,986,1044,492,493,494,1056,1048,1049,1057,1050,1052,1059,627,628,993,738,777,1042,960,815,1054,1043,1058],"class_list":["post-4972","post","type-post","status-publish","format-standard","hentry","category-uncategorized","tag-ai","tag-aisb","tag-alan-turing","tag-artificial-intelligence","tag-autonomous-created-intelligence","tag-autonomous-creative-intelligence","tag-autonomous-generated-intelligence","tag-autonomously-creative-intelligence","tag-blade-runner","tag-cautionary-tales","tag-david-j-gunkel","tag-ethics","tag-fairytales","tag-frankenstein","tag-iacap","tag-jamais-cascio","tag-joanna-j-bryson","tag-machine-consciousness","tag-machine-ethics","tag-machine-intelligence","tag-mary-shelley","tag-moral-agency","tag-moral-patiency","tag-neuromancer","tag-object-lessons","tag-parables","tag-peter-singer","tag-philosophy","tag-philosophy-of-mind","tag-science-fiction","tag-self-fulfilling-prophecy","tag-speculative-fiction","tag-splice","tag-surveillance-culture","tag-terminator-the-sarah-connor-chronicles","tag-the-golem-of-prague","tag-the-machine-question-symposium","tag-william-gibson"],"jetpack_publicize_connections":[],"jetpack_featured_media_url":"","jetpack_sharing_enabled":true,"jetpack_shortlink":"https:\/\/wp.me\/p5WByP-1ic","jetpack_likes_enabled":true,"jetpack-related-posts":[{"id":5316,"url":"https:\/\/afutureworththinkingabout.com\/?p=5316","url_meta":{"origin":4972,"position":0},"title":"My Appearance on The Machine Ethics Podcast&#8217;s A.I. Retreat Episode","author":"Damien P. Williams","date":"October 23, 2018","format":false,"excerpt":"As you already know, we went to the second Juvet A.I. Retreat, back in September. If you want to hear several of us talk about what we got up to at the then you're in luck because here are several conversations conducted by Ben Byford of the Machine Ethics Podcast.\u2026","rel":"","context":"In \"algorithmic bias\"","block_context":{"text":"algorithmic bias","link":"https:\/\/afutureworththinkingabout.com\/?tag=algorithmic-bias"},"img":{"alt_text":"","src":"https:\/\/i0.wp.com\/img.youtube.com\/vi\/ownE2zxTN2U\/0.jpg?resize=350%2C200","width":350,"height":200},"classes":[]},{"id":4966,"url":"https:\/\/afutureworththinkingabout.com\/?p=4966","url_meta":{"origin":4972,"position":1},"title":"BBC: &#8220;Tech giants pledge $1bn for &#8216;altruistic AI&#8217; venture, OpenAI&#8221;","author":"Damien P. Williams","date":"December 12, 2015","format":false,"excerpt":"This headline comes from a piece over at the BBC that opens as follows: Prominent tech executives have pledged $1bn (\u00a3659m) for OpenAI, a non-profit venture that aims to develop artificial intelligence (AI) to benefit humanity. The venture's backers include Tesla Motors and SpaceX CEO Elon Musk, Paypal co-founder Peter\u2026","rel":"","context":"In \"A Future Worth Thinking About\"","block_context":{"text":"A Future Worth Thinking About","link":"https:\/\/afutureworththinkingabout.com\/?tag=a-future-worth-thinking-about"},"img":{"alt_text":"","src":"","width":0,"height":0},"classes":[]},{"id":1185,"url":"https:\/\/afutureworththinkingabout.com\/?p=1185","url_meta":{"origin":4972,"position":2},"title":"Hey everyone. As you should\u2026","author":"Damien P. Williams","date":"February 8, 2015","format":false,"excerpt":"Hey everyone. As you should be aware, by now, there's the new WordPress blog for text posts: http:\/\/afutureworththinkingabout.wordpress.com So I'll be spending the next few days transferring older text posts from here, to there. Woooooo. Tell your friends. ;)","rel":"","context":"In \"A Future Worth Thinking About\"","block_context":{"text":"A Future Worth Thinking About","link":"https:\/\/afutureworththinkingabout.com\/?tag=a-future-worth-thinking-about"},"img":{"alt_text":"","src":"","width":0,"height":0},"classes":[]},{"id":4859,"url":"https:\/\/afutureworththinkingabout.com\/?p=4859","url_meta":{"origin":4972,"position":3},"title":"My First Appearance on Mindful Cyborgs","author":"Damien P. Williams","date":"April 29, 2015","format":false,"excerpt":"I sat down with Klint Finley of\u00a0Mindful Cyborgs to talk about many, many things: \u2026pop culture portrayals of human enhancement and artificial intelligence and why we need to craft more nuanced narratives to explore these topics\u2026 Tune in next week to hear Damien talk about how AI and transhumanism intersects\u2026","rel":"","context":"In \"A Future Worth Thinking About\"","block_context":{"text":"A Future Worth Thinking About","link":"https:\/\/afutureworththinkingabout.com\/?tag=a-future-worth-thinking-about"},"img":{"alt_text":"","src":"","width":0,"height":0},"classes":[]},{"id":5039,"url":"https:\/\/afutureworththinkingabout.com\/?p=5039","url_meta":{"origin":4972,"position":4},"title":"Direct Link For &#8220;The Metaphysical Cyborg&#8221;","author":"Damien P. Williams","date":"July 31, 2016","format":false,"excerpt":"Here's the direct link to my paper 'The Metaphysical Cyborg' from Laval Virtual 2013. Here's the abstract: \"In this brief essay, we discuss the nature of the kinds of conceptual changes which will be necessary to bridge the divide between humanity and machine intelligences. From cultural shifts to biotechnological integration,\u2026","rel":"","context":"In \"artificial intelligence\"","block_context":{"text":"artificial intelligence","link":"https:\/\/afutureworththinkingabout.com\/?tag=artificial-intelligence"},"img":{"alt_text":"","src":"","width":0,"height":0},"classes":[]},{"id":5082,"url":"https:\/\/afutureworththinkingabout.com\/?p=5082","url_meta":{"origin":4972,"position":5},"title":"From WIRED: &#8220;Tech Giants Team Up to Keep AI From Getting Out of Hand&#8221;","author":"Damien P. Williams","date":"September 28, 2016","format":false,"excerpt":"I spoke with Klint Finley over at WIRED about Amazon, Facebook, Google, IBM, and Microsoft's new joint ethics and oversight venture, which they've dubbed the \"Partnership on Artificial Intelligence to Benefit People and Society.\" They held a joint press briefing, today, in which Yann LeCun, Facebook's director of AI, and\u2026","rel":"","context":"In \"A Future Worth Thinking About\"","block_context":{"text":"A Future Worth Thinking About","link":"https:\/\/afutureworththinkingabout.com\/?tag=a-future-worth-thinking-about"},"img":{"alt_text":"","src":"","width":0,"height":0},"classes":[]}],"_links":{"self":[{"href":"https:\/\/afutureworththinkingabout.com\/index.php?rest_route=\/wp\/v2\/posts\/4972","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/afutureworththinkingabout.com\/index.php?rest_route=\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/afutureworththinkingabout.com\/index.php?rest_route=\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/afutureworththinkingabout.com\/index.php?rest_route=\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/afutureworththinkingabout.com\/index.php?rest_route=%2Fwp%2Fv2%2Fcomments&post=4972"}],"version-history":[{"count":7,"href":"https:\/\/afutureworththinkingabout.com\/index.php?rest_route=\/wp\/v2\/posts\/4972\/revisions"}],"predecessor-version":[{"id":5692,"href":"https:\/\/afutureworththinkingabout.com\/index.php?rest_route=\/wp\/v2\/posts\/4972\/revisions\/5692"}],"wp:attachment":[{"href":"https:\/\/afutureworththinkingabout.com\/index.php?rest_route=%2Fwp%2Fv2%2Fmedia&parent=4972"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/afutureworththinkingabout.com\/index.php?rest_route=%2Fwp%2Fv2%2Fcategories&post=4972"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/afutureworththinkingabout.com\/index.php?rest_route=%2Fwp%2Fv2%2Ftags&post=4972"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}