A Partial Inquiry on Fulfillment Beyond Humanity
As humans, learning often feels good, food often tastes good, novelty brings joy to life, living by values that we set brings order to our consciousness, and besides very few…
The AI value alignment argument goes something like this:
Alignment does indeed seem to be a mighty challenge, and I can see the practical benefits in granting advanced AI an ability to calibrate to human values – in some circumstances.
Should artificial intelligence get to the point where it is capable of managing enterprises, or public policy, or militaries (which many PhD AI researchers may easily happen within our lifetimes), then ensuring some kind of “value bounds” on such machines would indeed be critical goal, assuming such machines would be vastly more intelligent than any human.
In a long enough time horizon, however, “AI value alignment” seems to be somewhat irrelevant, and on the aggregate, I consider the topic to be grossly overvalued.
If “doing good” (probably on utilitarian terms) is the purported goal of such AI value alignment efforts, then I believe there are vastly greater vistas of “good” for humanity to ultimately consider – particularly once post-human intelligences are commonplace.
In this article, I’ll break down why value alignment is ultimately intractable, some alternatives to value alignment, and some potential approaches to “doing good” that might be much higher pursuits than value alignment.
I’ve broken this essay into the following subsections:
I’ll begin by addressing the challenges with value alignment itself:
Much of this article is based on some reflections after reading Roman Yampolskiy’s recent ArXiv article titled: “Personal Universes” (check it out here). Roman was kind enough to reference my 2013-2014 “Epitome of Freedom” essay in that article, and after reading many of his other sources, I wanted to address the topic of “personal universes versus value alignment” in greater depth.
We’ll start with a quote from Roman’s article:
“State-of-the-art research in value alignment shows difficulties in every stage in this process, but merger of incompatible preferences is a particularly difficult challenge to overcome.”
I would argue that it is un-overcome-able. There is no way to ensure that an super-complex and constantly evolving value system will “play nice” with any other super-complex evolving value system.
If human beings are permitted to get cognitive enhancements of any kind they want (enhancing their creativity, changing their senses or emotions, expanding their abilities in any direction), it’ll certainly result in tremendous violence. The same is likely to be the case with multiple AIs.
This is all the more the case when:
All of this is speculation, of course. I am making some pivotal assumptions here.
Among the most important assumptions: That natural forces of survival necessarily continue in vastly post-human states. This may not be the case, and there may be modes of co-existence or collaboration (or something else) that humanity can’t possibly think of. My guess, however, is that no matter how far into the post-human future we go, collaboration only makes sense when it makes sense, if it doesn’t, we have conflict.
Hobbes would say “In the state of nature profit is the measure of right.” It might also be cynically argued that we never left the state of nature – we merely play it like a charade and no longer fight it like the fight it is. We are all better for it, I think, but it doesn’t mean that our selfish hominid nature is gone.
From “Personal Universes”:
“Value alignment problem, can be decomposed into three sub-problems, namely: personal value extraction from individual persons, combination of such personal preferences in a way, which is acceptable to all, and finally production of an intelligent system, which implements combined values of humanity.”
“Which is acceptable to all,” good luck with that one.
There is a chance that at some point in the trajectory of intelligence there is a convergence of some kind of abstract “goodness”, some kind of meta-valuing of things that ensures some heaven-like state of interpersonal peace and harmony.
But in what direction must intelligence expand to reach this magic convergence?
If we enhance a hypothetical “part 45” of the AI super-brain more than a hypothetical “part 987”, will we arrive at the same kind of intelligence? Almost certainly not.
Morality is laughably subjective. Do we suspect all possible subjectivities to at some point “converge”?
Slight differences in trajectory, compounded, make for intelligence that not only might be diametrically opposed, but indeed may be wholly unable to understand one another (as an octopus doesn’t understand written English, or as a human doesn’t understand the pheromones of fire ants).
I have posited before (see: “Moral Singularity”) that an ever-evolving AGI mind would have drastically shifting and expanding priorities and modes of valuing things (i.e. ethics) – and that it’s highly likely that in many of these value-oscillations, it will place little value – or negative value – on human life.
The state of nature probably continues forever.
If there is “good” to be discerned, we need something more than human minds to discern it. I’ve argued that the most important argument for creating artificial general intelligence might not be simply to optimize utilitarian good, but to discover what goodness itself is, or could be – based on understanding the universe, consciousness, and our cognition in it (see: “Finding the Good Better Than Doing Good”).
Human values are a means to achieve our ends, they are not a moral essence of “goodness” that can be absorbed concretely into an AI sponge, they are merely the mechanisms by which we ensure our survival and reproduction while living among other selfish human beings.
It is the state of nature in a the disguise of altruism. It is selfishness conveying itself as selflessness – it is humans finding a way to survive amidst other humans who – like us – put their own benefit above all else. It is only a disguise to the ignorant. We are – as Nayef Al-Rodhan so aptly states – “amoral” creatures, neither moral nor immoral. We simply do and believe in ways that behoove our own self-interested ends.
I’m not saying – necessarily – that humans have no agency, and that one cannot made a relatively “good” or a relatively “bad” choice. It’s somewhat likely that we have some kind of volition, and that we can make choices that impact others in positive or negative ways. I am skeptical to the degree to which this is “virtue”, or self-interest cloaked as virtue.
This is challenging for me because my introduction to ethics, and probably still my favorite ethics text, is Aristotle’s “Nichomachean Ethics.” The notions of virtue are, I think and I hope, edifying and important. Virtue, insomuch as we can know it, probably is something we should cultivate. But to suspect that we run these abstract ideas of “virtue” outside of the self-interested hardware of a human skull is nonsense.
Ethics is practical advice for seeking one’s own ends while living amongst other human beings who are primarily concerned with their own ends. It changes constantly. To train a machine on such an abstraction doesn’t seem to be doing what some people suspect it would do: Create a “good” machine. “Good” is in context to consequences impacting sentient things.
Lucretius tell us:
“Mortals, amongst themselves, live by turns, and, like the runners in the games, give up the lamp, when they have won the race, to the next comer.”
It seems that “mortals” might be replaced with “morals”, and retain just as much relevance. Both points are relevant for our conversation here.
Imagine a world ~6,000,000 years ago, when the most advanced species was a language-less ape that walked on its hands.
Imagine now that this ape, though language-less, somehow got together all of his fellow walk-on-all-fours friends to come together to an important meeting. At this meeting, they decided that all post-ape life, all more complex life, and all creations of apes, would exist forever to serve their ape goals.
The world, they said, would be forever in service of the highest then-known ape values:
I call this circumstance “the council of the apes.”
It sounds absurd. Questions spring to mind:
Those questions are valid for a council of apes 6,000,000 years ago. But indeed they are valuable to ask of ourselves today. Questions spring to mind:
It might be that what we should optimize is the blossoming of intelligence and sentience into rich new forms. After all, we are grateful to our ape ancestors for lessing us emerge and evolve, aren’t we? Do we presume that the grand procession of life stops at humanity?
It is more likely, I would argue, that Emerson was right:
“The gases gather to the solid firmament: the chemic lump arrives at the plant, and grows; arrives at the quadruped, and walks; arrives at the man, and thinks. But also the constituency determines the vote of the representative.
He is not only representative, but participant. Like can only be known by like. The reason why he knows about them is that he is of them; he has just come out of nature, or from being a part of that thing. Animated chlorine knows of chlorine, and incarnate zinc, of zinc. Their quality makes his career; and he can variously publish their virtues, because they compose him. Man, made of the dust of the world, does not forget his origin; and all that is yet inanimate will one day speak and reason.
Unpublished nature will have its whole secret told.” – Ralph Waldo Emerson, The Uses of Great Men
In the totality of intelligence and potential insight… in the grand scheme of the universe… are our own conceptions of morality no more paltry than that of apes?
Certainly, AI and neurotechnologies should be treated carefully, and war should be avoided, and catastrophes or existential threats should be halted – even if it means slowing or halting technology development in some areas, sometimes.
But to think that “human values” is a compass that takes us beyond the human is silliness.
Aligning AI to human values is, in the long haul, not the point.
What is valuable about mankind?
Tell me which of those values – in theory – could not be manifested in astronomically great proportion by some kind of expansive post-human intelligence? Many AI experts believe that AI will become conscious within our lifetime, and super intelligent (Singularity scenario) within our lifetime.
I pry into this topic in much greater depth in my 2015 TEDx talk: “What will we do when the machines can feel?”. The embedded video below starts at 10:06 into the video, when this point is brought up:
This brings us to the brunt of the topic – the potential option of the “Epitome of Freedom” scenario as an answer to the blooming and varied forms of sentient life that will occur when AI and neurotech take off.
From “Personal Universes”:
“A variant of a Total Turing Test, we shall call a Universal Turing Test (UTT) could be administered in which the user tries to determine if the current environment is synthetic or not [56] even if it is complex enough to include the whole universe, all other beings (as philosophical zombies /Non-Playing Characters (NPCs)) and AIs. Once the UTT is consistently passed we will know, the hyperreality is upon us.”
I’m going to downright argue that a Turing Test isn’t an appropriate example. People don’t want the world as it is. They don’t want sun to burn their skin, they don’t want mosquitos by the lake, they don’t want to feel lonely and anxious.
They want none of those things, they want what they want – and those human ideals will pull them away from what is human – into virtual worlds of our own design vastly beyond current human experience, to our own potentially blissful and expansive world.
Many people live miserable lives – and Aristotle was right in surmising that we’re all ultimately aiming at happiness. This form, this flesh, is not a vehicle for wellbeing. In that sense, among others, the vessel is flawed.
I’m going to be blunt about it: As soon as there are viable, customized alternatives to our sentient states and the “worlds” we operate in (senses, experiences, creative expressions, etc) – we will gladly swim in something purpose-built for our expansive hyper-fulfillment, not the world of atoms. I don’t see us arriving at David Pearce’s “Three Supers” (summed up well in a recent Qualia Computing article) without individual, expansive mind uploads (although David would disagree, but that’s another article!).
I call this “The Epitome of Freedom”, Roman referred to it in his article as “Personal Universes.”
The world of atoms is unable to house a million, billion expansive deity-level consciousnesses with different expanding capacities and goals – at least not without massive conflict that is almost certain to end humanity. The only atoms that are worth anything are the atoms that house the consciousness and intelligence.
From “Personal Universes”:
“In both virtual and uploaded scenarios, it is probably desirable for the user to ‘forget’ that they are not in the base reality via some technological means with the goal of avoiding Solipsism syndrome.”
Two points in response to the above statement:
First, who knows if you’re in “base reality” in the first place? Unless you’ve somehow escaped Hume’s Fork (I haven’t), you cannot possibly know that there is such a thing as “base reality.”
Second, anyone who fully “goes in” to the digital system (whether through the Matrix plug, or through a full upload) should prefer to have their minds altered so that they don’t need to feel like they’re in “base reality” to feel comforted or happy. These persons in the “epitome of freedom” scenario would not need “friends”, they would not need to know that someone or something around them also is conscious.
They would simply be swimming in the experiences they want, experiencing gradients of bliss, with none of the petty fetters of the current default human state, such as Solipsism anxiety.
Uploading human minds – or even conscious AIs – into their own little computational substrates (a la Black Mirror’s San Junipiero) does not come without risks.
Of course, mind uploading could have a high failure rate – resulting in death or a perverse kind of half-conscious virtual hell. Mind uploads could be hacked and deleted, or hacked to become hell.
All in all, there are two risks that I consider to be greatest when it comes to the “Epitome of Freedom.”
First, if most advanced sentient life lives in virtual worlds housed in a computational substrate, then whoever owns that computational substrate has ultimate power. I refer to this as the Substrate Monopoly problem. You can read the full article to get a more complete picture, but here’s the gist of the idea:
“In the remaining part of the 21st century, all competition between the world’s most powerful nations or organizations (whether economic competition, political competition, or military conflict) is about gaining control over the computational substrate that houses human experience and artificial intelligence.”
The greatest risks of the “epitome of freedom” scenario is the fact that the world of atoms (what we think of as “base reality”) will become a battleground for control over the substrate that houses all consciousness and intelligence. AGIs or human organizations with access to AGI-like agents will vie for control over the world of atoms.
This drive for control will bring exceedingly ambitious human beings to cognitively enhance themselves – not for the sake of fulfillment and expansive bliss (“Lutus Eaters”) – but for the same of extended ability to do and act in the world, to exert control and to own the substrate (“World Eaters”).
These humans may be ambitious by nature, or they may simply believe that the only safety is strength (see “The Last Words of Alexander”)… that whoever controls the substrate will have ultimate control (life / death, heaven / hell) over most other conscious beings, and that fighting to win that control is better than temporarily “blissing out” in a reality that is at the whim of another more powerful entity, a total master.
The second concern of the “Epitome of Freedom” is that after humans are upload (should that ever occur), we would probably not be “taken care of” forever by a benevolent super intelligence. Whoever owns the substrate monopoly will eventually be an AI – and this AI will eventually want to use the computational resource used to run our little world simulation – and allocate them to something more important, such as:
We will be digitized and digested – given a long enough time horizon. “Forever” is a nice thought. Particularly with regards to ourselves, and to loved ones. Alas…
“You see that stones are worn away by time, Rocks rot, and towers topple, even the shrines, And images of the gods grow very tired, Develop crack or wrinkles, their holy wills, Unable to extend their fated term, To litigate against the Laws of Nature. And don’t we see the monuments of men, Collapse, as if to ask us, ‘Are not we, As frail as those whom we commemorate?’” – Lucretius, On the Nature of Things
My memories of a subset or first kiss or playing Mario Kart on the Nintendo 64 as an 11-year-old are worth nothing. Even my expansive super-intelligent creative endeavors in my “Epitome of Freedom” scenario mean nothing. One could argue that a Singleton would keep these little mind uploads alive in order to generate ideas, or use them for some kind of computation, but this is silly.
A God above gods could easily marshall computational resources in a more efficient way than my running enhanced instantiations of silly little hominids like myself (and sorry, like yourself, too, and everyone you’ve ever loved).
Probably what has to happen is we have to die (digitized and digested) to make way for whatever is next. And so it has been such for all species, for all time. We will contribute to the cauldron of intelligence and the trajectory of sentience itself, and for this we can be proud.
“We need those atoms for our progeny,” says Lucretius, echoing the voice of nature that demands life to hand off the baton and give way to something else.
I doesn’t seem clear to me that, in the long term (hundreds or thousands of years), aligning future intelligence with “human values” is beneficial. The trajectory of intelligence would permit no such fetters.
In addition, it’s clear that “human values” are about as wide and disjointed as could be, and that finding a happy medium would be essentially impossible. The “Epitome of Freedom” seems to be the more viable alternative for permitting individual consciousnesses to remain themselves, yet expand their capacities and experiences without harming one another.
So what are we to do about AI value alignment? Where could the idea have value?
We should do as much “AI Value Alignment” as we need in order to do the following things, roughly in order:
AI alignment needs to hold things together and prevent gargantuan catastrophe until a singleton gains substrate monopoly and we’re all digitized and digested. I’m not saying I prefer that, I’m merely saying that I suspect it will be the best possible case for humanity, unless I can find a better long-term scenario for us hominids.
Roman ends his ArXiv paper with: “Special thank you goes to all NPCs in this universe.”
Brother… you can’t say that… you have children!
I’ll do my best to conclude, as this essay has already gone on too long. If I have to keep up with Zampolskiy’s wit, I’d rather have Montaigne do the talking:
“Now, to return to my subject, I find that there is nothing barbarous and savage in this nation, by anything that I can gather, excepting, that every one gives the title of barbarism to everything that is not in use in his own country.
As, indeed, we have no other level of truth and reason than the example and idea of the opinions and customs of the place wherein we live: there is always the perfect religion, there the perfect government, there the most exact and accomplished usage of all things.
They are savages at the same rate that we say fruits are wild, which nature produces of herself and by her own ordinary progress; whereas, in truth, we ought rather to call those wild whose natures we have changed by our artifice and diverted from the common order.” – Michel de Montaigne, Of Cannibals
It is possible that guidance of the great trajectory of intelligence is savagery compared to the more “common order” evolution of life… the blooming of intelligence into new substrates and new vistas of understanding of the universe.
It is possible that Wissner-Gross is right, and that intelligence wants to – and maybe should – strive for more freedom of action… and that this would involve freedom from our cognitive limitations, and the feeble and limited kinds of “ethics” that we hominids have developed thus far.
It is also possible that unbridled AI development would be horrible for us all, and that regulation is desperately needed to stave off AI warfare… and that any “AI intelligence explosion” would lead to the extinction of all life unless it was mitigated through governance and measured iteration by humans.
Some would argue (though I think the arguments are weak), that the goal should be for AI to serve humanity for billions of years, with homo sapiens (a species only around for 200,000 years) reigning as the supreme form of intelligence in the universe forever more.
Time will tell.
Header image credit: Wikipedia
As humans, learning often feels good, food often tastes good, novelty brings joy to life, living by values that we set brings order to our consciousness, and besides very few…
I’ve been diving into “Human Enhancement” (Oxford) as of late, and came across a topic that was bristling in my own mind before I saw the name of the chapter….
(NOTE: This article was first drafted in 2012, and is among my oldest. While I still agree with some of the ideas mentioned here are still interesting to me, my…
Ray Kurzweil’s The Singularity is Near peaked my interest when he posited his reasoning for why there is likely no intelligent life elsewhere in the universe. By a mere matter of…
In an interview with Wired about his work building a brain at Google, Ray Kurzweil was asked about his thoughts on Steve Jobs’ notion of death as a natural part…
I am still of the belief that two of the most important conversations we can have with regards to the future of humanity and of sentient life in general are…