Commit
·
b22086d
1
Parent(s):
cfc64ca
Upload experiment.txt
Browse files- experiment.txt +173 -0
experiment.txt
ADDED
@@ -0,0 +1,173 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
1. What Is The Singularity?
|
2 |
+
The acceleration of technological progress has been the central feature of this century. We are on the edge of change comparable to the rise of human life on Earth. The precise cause of this change is the imminent creation by technology of entities with greater-than-human intelligence. Science may achieve this breakthrough by several means (and this is another reason for having confidence that the event will occur):
|
3 |
+
|
4 |
+
Computers that are "awake" and superhumanly intelligent may be developed. (To date, there has been much controversy as to whether we can create human equivalence in a machine. But if the answer is "yes," then there is little doubt that more intelligent beings can be constructed shortly thereafter.)
|
5 |
+
|
6 |
+
Large computer networks and their associated users may "wake up" as superhumanly intelligent entities.
|
7 |
+
|
8 |
+
Computer/human interfaces may become so intimate that users may reasonably be considered superhumanly intelligent.
|
9 |
+
|
10 |
+
Biological science may provide means to improve natural human intellect.
|
11 |
+
|
12 |
+
The first three possibilities depend on improvements in computer hardware. Progress in hardware has followed an amazingly steady curve in the last few decades. Based on this trend, I believe that the creation of greater-than-human intelligence will occur during the next thirty years. (Charles Platt has pointed out that AI enthusiasts have been making claims like this for thirty years. Just so I'm not guilty of a relative-time ambiguity, let me be more specific: I'll be surprised if this event occurs before 2005 or after 2030.)
|
13 |
+
|
14 |
+
What are the consequences of this event? When greater-than-human intelligence drives progress, that progress will be much more rapid. In fact, there seems no reason why progress itself would not involve the creation of still more intelligent entities -- on a still-shorter time scale. The best analogy I see is to the evolutionary past: Animals can adapt to problems and make inventions, but often no faster than natural selection can do its work -- the world acts as its own simulator in the case of natural selection. We humans have the ability to internalize the world and conduct what-if's in our heads; we can solve many problems thousands of times faster than natural selection could. Now, by creating the means to execute those simulations at much higher speeds, we are entering a regime as radically different from our human past as we humans are from the lower animals.
|
15 |
+
|
16 |
+
This change will be a throwing-away of all the human rules, perhaps in the blink of an eye -- an exponential runaway beyond any hope of control. Developments that were thought might only happen in "a million years" (if ever) will likely happen in the next century.
|
17 |
+
|
18 |
+
It's fair to call this event a singularity ("the Singularity" for the purposes of this piece). It is a point where our old models must be discarded and a new reality rules, a point that will loom vaster and vaster over human affairs until the notion becomes a commonplace. Yet when it finally happens, it may still be a great surprise and a greater unknown. In the 1950s very few saw it: Stan Ulam1 paraphrased John von Neumann as saying:
|
19 |
+
|
20 |
+
One conversation centered on the ever-accelerating progress of technology and changes in the mode of human life, which gives the appearance of approaching some essential singularity in the history of the race beyond which human affairs, as we know them, could not continue.
|
21 |
+
Von Neumann even uses the term singularity, though it appears he is thinking of normal progress, not the creation of superhuman intellect. (For me, the superhumanity is the essence of the Singularity. Without that we would get a glut of technical riches, never properly absorbed.)
|
22 |
+
|
23 |
+
The 1960s saw recognition of some of the implications of superhuman intelligence. I. J. Good2 wrote:
|
24 |
+
|
25 |
+
Let an ultraintelligent machine be defined as a machine that can far surpass all the intellectual activities of any man however clever. Since the design of machines is one of these intellectual activities, an ultraintelligent machine could design even better machines; there would then unquestionably be an "intelligence explosion," and the intelligence of man would be left far behind. Thus the first ultraintelligent machine is the last invention that man need ever make, provided that the machine is docile enough to tell us how to keep it under control. . . . It is more probable than not that, within the twentieth century, an ultraintelligent machine will be built and that it will be the last invention that man need make.
|
26 |
+
Good has captured the essence of the runaway, but he does not pursue its most disturbing consequences. Any intelligent machine of the sort he describes would not be humankind's "tool" -- any more than humans are the tools of rabbits, robins, or chimpanzees.
|
27 |
+
|
28 |
+
Through the sixties and seventies and eighties, recognition of the cataclysm spread. Perhaps it was the science-fiction writers who felt the first concrete impact. After all, the "hard" science-fiction writers are the ones who try to write specific stories about all that technology may do for us. More and more, these writers felt an opaque wall across the future. Once, they could put such fantasies millions of years in the future. Now they saw that their most diligent extrapolations resulted in the unknowable . . . soon. Once, galactic empires might have seemed a Posthuman domain. Now, sadly, even interplanetary ones are.
|
29 |
+
|
30 |
+
What about the coming decades, as we slide toward the edge? How will the approach of the Singularity spread across the human world view? For a while yet, the general critics of machine sapience will have good press. After all, until we have hardware as powerful as a human brain it is probably foolish to think we'll be able to create human-equivalent (or greater) intelligence. (There is the farfetched possibility that we could make a human equivalent out of less powerful hardware -- if we were willing to give up speed, if we were willing to settle for an artificial being that was literally slow. But it's much more likely that devising the software will be a tricky process, involving lots of false starts and experimentation. If so, then the arrival of self-aware machines will not happen until after the development of hardware that is substantially more powerful than humans' natural equipment.)
|
31 |
+
|
32 |
+
But as time passes, we should see more symptoms. The dilemma felt by science-fiction writers will be perceived in other creative endeavors. (I have heard thoughtful comicbook writers worry about how to create spectacular effects when everything visible can be produced by the technologically commonplace.) We will see automation replacing higher- and higher-level jobs. We have tools right now (symbolic math programs, cad/cam) that release us from most low-level drudgery. Put another way: the work that is truly productive is the domain of a steadily smaller and more elite fraction of humanity. In the coming of the Singularity, we will see the predictions of true technological unemployment finally come true.
|
33 |
+
|
34 |
+
Another symptom of progress toward the Singularity: ideas themselves should spread ever faster, and even the most radical will quickly become commonplace.
|
35 |
+
|
36 |
+
And what of the arrival of the Singularity itself? What can be said of its actual appearance? Since it involves an intellectual runaway, it will probably occur faster than any technical revolution seen so far. The precipitating event will likely be unexpected -- perhaps even by the researchers involved ("But all our previous models were catatonic! We were just tweaking some parameters . . ."). If networking is widespread enough (into ubiquitous embedded systems), it may seem as if our artifacts as a whole had suddenly awakened.
|
37 |
+
|
38 |
+
And what happens a month or two (or a day or two) after that? I have only analogies to point to: The rise of humankind. We will be in the Posthuman era. And for all my technological optimism, I think I'd be more comfortable if I were regarding these transcendental events from one thousand years' remove . . . instead of twenty.
|
39 |
+
|
40 |
+
2. Can the Singularity Be Avoided?
|
41 |
+
Well, maybe it won't happen at all: sometimes I try to imagine the symptoms we should expect to see if the Singularity is not to develop. There are the widely respected arguments of Penrose3 and Searle4 against the practicality of machine sapience. In August 1992, Thinking Machines Corporation held a workshop to investigate "How We Will Build a Machine That Thinks." As you might guess from the workshop's title, the participants were not especially supportive of the arguments against machine intelligence. In fact, there was general agreement that minds can exist on nonbiological substrates and that algorithms are of central importance to the existence of minds. However, there was much debate about the raw hardware power present in organic brains. A minority felt that the largest 1992 computers were within three orders of magnitude of the power of the human brain. The majority of the participants agreed with Hans Moravec's estimate5 that we are ten to forty years away from hardware parity. And yet there was an other minority who conjectured that the computational competence of single neurons may be far higher than generally believed. If so, our present computer hardware might be as much as ten orders of magnitude short of the equipment we carry around in our heads. If this is true (or for that matter, if the Penrose or Searle critique is valid), we might never see a Singularity. Instead, in the early '00s we would find our hardware performance curves beginning to level off -- because of our inability to automate the design work needed to support further hardware improvements. We'd end up with some very powerful hardware, but without the ability to push it further. Commercial digital signal processing might be awesome, giving an analog appearance even to digital operations, but nothing would ever "wake up" and there would never be the intellectual runaway that is the essence of the Singularity. It would likely be seen as a golden age . . . and it would also be an end of progress. This is very like the future predic ted by Gunther Stent,6 who explicitly cites the development of transhuman intelligence as a sufficient condition to break his projections.
|
42 |
+
|
43 |
+
But if the technological Singularity can happen, it will. Even if all the governments of the world were to understand the "threat" and be in deadly fear of it, progress toward the goal would continue. The competitive advantage -- economic, military, even artistic -- of every advance in automation is so compelling that forbidding such things merely assures that someone else will get them first.
|
44 |
+
|
45 |
+
Eric Drexler has provided spectacular insights about how far technical improvement may go.7 He agrees that superhuman intelligences will be available in the near future. But Drexler argues that we can confine such transhuman devices so that their results can be examined and used safely.
|
46 |
+
|
47 |
+
I argue that confinement is intrinsically impractical. Imagine yourself locked in your home with only limited data access to the outside, to your masters. If those masters thought at a rate -- say -- one million times slower than you, there is little doubt that over a period of years (your time) you could come up with a way to escape. I call this "fast thinking" form of superintelligence "weak superhumanity." Such a "weakly superhuman" entity would probably burn out in a few weeks of outside time. "Strong superhumanity" would be more than cranking up the clock speed on a human-equivalent mind. It's hard to say precisely what "strong superhumanity" would be like, but the difference appears to be profound. Imagine running a dog mind at very high speed. Would a thousand years of doggy living add up to any human insight? Many speculations about superintelligence seem to be based on the weakly superhuman model. I believe that our best guesses about the post-Singularity world can be obtained by thinking on the nat ure of strong superhumanity. I will return to this point.
|
48 |
+
|
49 |
+
Another approach to confinement is to build rules into the mind of the created superhuman entity. I think that any rules strict enough to be effective would also produce a device whose ability was clearly inferior to the unfettered versions (so human competition would favor the development of the more dangerous models).
|
50 |
+
|
51 |
+
If the Singularity can not be prevented or confined, just how bad could the Posthuman era be? Well . . . pretty bad. The physical extinction of the human race is one possibility. (Or, as Eric Drexler put it of nanotechnology: given all that such technology can do, perhaps governments would simply decide that they no longer need citizens.) Yet physical extinction may not be the scariest possibility. Think of the different ways we relate to animals. A Posthuman world would still have plenty of niches where human-equivalent automation would be desirable: embedded systems in autonomous devices, self-aware daemons in the lower functioning of larger sentients. (A strongly superhuman intelligence would likely be a Society of Mind8 with some very competent components.) Some of these human equivalents might be used for nothing more than digital signal processing. Others might be very humanlike, yet with a onesidedness, a dedication that would put them in a mental hospital in our era. Though none of these creatures mi ght be flesh-and-blood humans, they might be the closest things in the new environment to what we call human now.
|
52 |
+
|
53 |
+
I have argued above that we cannot prevent the Singularity, that its coming is an inevitable consequence of humans' natural competitiveness and the possibilities inherent in technology. And yet: we are the initiators. Even the largest avalanche is triggered by small things. We have the freedom to establish initial conditions, to make things happen in ways that are less inimical than others. Of course (as with starting avalanches), it may not be clear what the right guiding nudge really is:
|
54 |
+
|
55 |
+
3. Other Paths to the Singularity
|
56 |
+
When people speak of creating superhumanly intelligent beings, they are usually imagining an AI project. But as I noted at the beginning of this article, there are other paths to superhumanity. Computer networks and human-computer interfaces seem more mundane than AI, yet they could lead to the Singularity. I call this contrasting approach Intelligence Amplification (IA). IA is proceeding very naturally, in most cases not even recognized for what it is by its developers. But every time our ability to access information and to communicate it to others is improved, in some sense we have achieved an increase over natural intelligence. Even now, the team of a Ph.D. human and good computer workstation (even an off-net workstation) could probably max any written intelligence test in existence.
|
57 |
+
|
58 |
+
And it's very likely that IA is a much easier road to the achievement of superhumanity than pure AI. In humans, the hardest development problems have already been solved. Building up from within ourselves ought to be easier than figuring out what we really are and then building machines that are all of that. And there is at least conjectural precedent for this approach. Cairns-Smith9 has speculated that biological life may have begun as an adjunct to still more primitive life based on crystalline growth. Lynn Margulis (in 10 and elsewhere) has made strong arguments that mutualism is a great driving force in evolution.
|
59 |
+
|
60 |
+
Note that I am not proposing that AI research be ignored. AI advances will often have applications in IA, and vice versa. I am suggesting that we recognize that in network and interface research there is something as profound (and potentially wild) as artificial intelligence. With that insight, we may see projects that are not as directly applicable as conventional interface and network design work, but which serve to advance us toward the Singularity along the IA path.
|
61 |
+
|
62 |
+
Here are some possible projects that take on special significance, given the IA point of view:
|
63 |
+
|
64 |
+
Human/computer team automation: Take problems that are normally considered for purely machine solution (like hillclimbing problems), and design programs and interfaces that take advantage of humans' intuition and available computer hardware. Considering the bizarreness of higher-dimensional hillclimbing problems (and the neat algorithms that have been devised for their solution), some very interesting displays and control tools could be provided to the human team member.
|
65 |
+
|
66 |
+
Human/computer symbiosis in art: Combine the graphic generation capability of modern machines and the esthetic sensibility of humans. Of course, an enormous amount of research has gone into designing computer aids for artists. I'm suggesting that we explicitly aim for a greater merging of competence, that we explicitly recognize the cooperative approach that is possible. Karl Sims has done wonderful work in this direction.11
|
67 |
+
|
68 |
+
Human/computer teams at chess tournaments: We already have programs that can play better than almost all humans. But how much work has been done on how this power could be used by a human, to get something even better? If such teams were allowed in at least some chess tournaments, it could have the positive effect on IA research that allowing computers in tournaments had for the corresponding niche in AI.
|
69 |
+
|
70 |
+
Interfaces that allow computer and network access without requiring the human to be tied to one spot, sitting in front of a computer. (This aspect of IA fits so well with known economic advantages that lots of effort is already being spent on it.)
|
71 |
+
|
72 |
+
More symmetrical decision support systems. A popular research/product area in recent years has been decision support systems. This is a form of IA, but may be too focused on systems that are oracular. As much as the program giving the user information, there must be the idea of the user giving the program guidance.
|
73 |
+
|
74 |
+
Local area nets to make human teams more effective than their component members. This is generally the area of "groupware"; the change in viewpoint here would be to regard the group activity as a combination organism.
|
75 |
+
|
76 |
+
In one sense, this suggestion's goal might be to invent a "Rules of Order" for such combination operations. For instance, group focus might be more easily maintained than in classical meetings. Individual members' expertise could be isolated from ego issues so that the contribution of different members is focused on the team project. And of course shared databases could be used much more conveniently than in conventional committee operations.
|
77 |
+
|
78 |
+
The Internet as a combination human/machine tool. Of all the items on the list, progress in this is proceeding the fastest. The power and influence of the Internet are vastly underestimated. The very anarchy of the worldwide net's development is evidence of its potential. As connectivity, bandwidth, archive size, and computer speed all increase, we are seeing something like Lynn Margulis' vision of the biosphere as data processor recapitulated, but at a million times greater speed and with millions of humanly intelligent agents (ourselves).
|
79 |
+
|
80 |
+
The above examples illustrate research that can be done within the context of contemporary computer science departments. There are other paradigms. For example, much of the work in artificial intelligence and neural nets would benefit from a closer connection with biological life. Instead of simply trying to model and understand biological life with computers, research could be directed toward the creation of composite systems that rely on biological life for guidance, or for the features we don't understand well enough yet to implement in hardware. A longtime dream of science fiction has been direct brain-to-computer interfaces. In fact, concrete work is being done in this area:
|
81 |
+
|
82 |
+
Limb prosthetics is a topic of direct commercial applicability. Nerve-to-silicon transducers can be made. This is an exciting near-term step toward direct communication.
|
83 |
+
|
84 |
+
Direct links into brains seem feasible, if the bit rate is low: given human learning flexibility, the actual brain neuron targets might not have to be precisely selected. Even 100 bits per second would be of great use to stroke victims who would otherwise be confined to menu-driven interfaces.
|
85 |
+
|
86 |
+
Plugging into the optic trunk has the potential for bandwidths of 1 Mbit/second or so. But for this, we need to know the fine-scale architecture of vision, and we need to place an enormous web of electrodes with exquisite precision. If we want our high-bandwidth connection to add to the paths already present in the brain, the problem becomes vastly more intractable. Just sticking a grid of high-bandwidth receivers into a brain certainly won't do it. But suppose that the high-bandwidth grid were present as the brain structure was setting up, as the embryo developed. That suggests:
|
87 |
+
|
88 |
+
Animal embryo experiments. I wouldn't expect any IA success in the first years of such research, but giving developing brains access to complex simulated neural structures might, in the long run, produce animals with additional sense paths and interesting intellectual abilities.
|
89 |
+
|
90 |
+
I had hoped that this discussion of IA would yield some clearly safer approaches to the Singularity (after all, IA allows our participation in a kind of transcendence). Alas, about all I am sure of is that these proposals should be considered, that they may give us more options. But as for safety -- some of the suggestions are a little scary on their face. IA for individual humans creates a rather sinister elite. We humans have millions of years of evolutionary baggage that makes us regard competition in a deadly light. Much of that deadliness may not be necessary in today's world, one where losers take on the winners' tricks and are coopted into the winners' enterprises. A creature that was built de novo might possibly be a much more benign entity than one based on fang and talon.
|
91 |
+
|
92 |
+
The problem is not simply that the Singularity represents the passing of humankind from center stage, but that it contradicts our most deeply held notions of being. I think a closer look at the notion of strong superhumanity can show why that is.
|
93 |
+
|
94 |
+
4. Strong Superhumanity and the Best We Can Ask For
|
95 |
+
Suppose we could tailor the Singularity. Suppose we could attain our most extravagant hopes. What then would we ask for? That humans themselves would become their own successors, that whatever injustice occurred would be tempered by our knowledge of our roots. For those who remained unaltered, the goal would be benign treatment (perhaps even giving the stay-behinds the appearance of being masters of godlike slaves). It could be a golden age that also involved progress (leaping Stent's barrier). Immortality (or at least a lifetime as long as we can make the universe survive) would be achievable.
|
96 |
+
|
97 |
+
But in this brightest and kindest world, the philosophical problems themselves become intimidating. A mind that stays at the same capacity cannot live forever; after a few thousand years it would look more like a repeating tape loop than a person. To live indefinitely long, the mind itself must grow . . . and when it becomes great enough, and looks back . . . what fellow-feeling can it have with the soul that it was originally? The later being would be everything the original was, but vastly more. And so even for the individual, the Cairns-Smith or Lynn Margulis notion of new life growing incrementally out of the old must still be valid.
|
98 |
+
|
99 |
+
This "problem" about immortality comes up in much more direct ways. The notion of ego and self-awareness has been the bedrock of the hardheaded rationalism of the last few centuries. Yet now the notion of self-awareness is under attack from the artificial intelligence people. Intelligence Amplification undercuts our concept of ego from another direction. The post-Singularity world will involve extremely high-bandwidth networking. A central feature of strongly superhuman entities will likely be their ability to communicate at variable bandwidths, including ones far higher than speech or written messages. What happens when pieces of ego can be copied and merged, when self-awareness can grow or shrink to fit the nature of the problems under consideration? These are essential features of strong superhumanity and the Singularity. Thinking about them, one begins to feel how essentially strange and different the Posthuman era will be -- no matter how cleverly and benignly it is brought to be.
|
100 |
+
|
101 |
+
From one angle, the vision fits many of our happiest dreams: a time unending, where we can truly know one another and understand the deepest mysteries. From another angle, it's a lot like the worst-case scenario I imagined earlier.
|
102 |
+
|
103 |
+
In fact, I think the new era is simply too different to fit into the classical frame of good and evil. That frame is based on the idea of isolated, immutable minds connected by tenuous, low-bandwith links. But the post-Singularity world does fit with the larger tradition of change and cooperation that started long ago (perhaps even before the rise of biological life). I think certain notions of ethics would apply in such an era. Research into IA and high-bandwidth communications should improve this understanding. I see just the glimmerings of this now; perhaps there are rules for distinguishing self from others on the basis of bandwidth of connection. And while mind and self will be vastly more labile than in the past, much of what we value (knowledge, memory, thought) need never be lost. I think Freeman Dyson has it right when he says, "God is what mind becomes when it has passed beyond the scale of our comprehension."12
|
104 |
+
|
105 |
+
|
106 |
+
What Begins After the End of the Enlightenment?
|
107 |
+
Yuk Hui
|
108 |
+
Issue #96
|
109 |
+
January 2019
|
110 |
+
In June 2018, Henry Kissinger published an article in The Atlantic titled “How the Enlightenment Ends.” At first glance, the article seems to suggest that the Enlightenment, the “Age of Reason,” has been put to an end by artificial intelligence. Machines that have the capacity for analysis and reasoning are overtaking human cognitive capacity. Technology that is rooted in Enlightenment thought is superseding the philosophy that is its fundamental principle. In view of this end of the Enlightenment, Kissinger proposes that it is necessary to look for a new philosophy: “The Enlightenment started with essentially philosophical insights spread by a new technology. Our period is moving in the opposite direction. It has generated a potentially dominating technology in search of a guiding philosophy.”1
|
111 |
+
|
112 |
+
However, we must first of all ask: Why does the increasing capacity of machines necessarily mean the end of the Enlightenment? And why does it lead the former US Secretary of State, at the end of his article, to call for the US to prioritize artificial intelligence research as a matter of immediate national concern?
|
113 |
+
|
114 |
+
If we begin with Kissinger’s title, “How the Enlightenment Ends,” we might ask how an “unfinished project” (in the words of Jürgen Habermas) such as the Enlightenment can be finished at all.2 Or has the China expert now joined the anti-Enlightenment tradition of Giambattista Vico, Johann Gottfried von Herder, Edmund Burke, Thomas Carlyle, Hippolyte Taine, Ernest Renan, Benedetto Croce, Friedrich Meinecke, Oswald Spengler, sometimes Nietzsche, and, more recently, Nick Land, another China expert? What follows is my response to Kissinger’s article; it can also be read as a continuation of two essays I published previously in e-flux journal, “On the Unhappy Consciousness of Neoreactionaries” and “Cosmotechnics as Cosmopolitics.”3
|
115 |
+
|
116 |
+
|
117 |
+
1. The “Decisive Mistakes” of “the ‘White’ Peoples”
|
118 |
+
Happy progressivists have followed the emanation of the Enlightenment to its end and found that, in fact, the light leads to total darkness. The end comes as a huge surprise: the void. Are Kissinger’s words not witness to this end as the abyss of humanity? But when the Secretary of State asks computer scientists to understand the history of philosophy, he does so without specifying which history and which philosophy. Kissinger describes the suffering that ensues when the high point of Western civilization has passed; Oswald Spengler called this the “decline” or “downfall” (Untergang) of the West. Similarities between the ideas of Kissinger and Spengler are far from coincidental, not least because the latter was the subject of the former’s honors thesis at Harvard. Titled “The Meaning of History: Reflections on Spengler, Toynbee, and Kant,” the thesis focused on determinism and freedom in history, following from Spengler’s description of history as an organic process. As Kissinger wrote: “Life is suffering, birth involves death. Transitoriness is the fate of existence. No civilization has yet been permanent, no longing completely fulfilled. This is necessity, the fatedness of history, the dilemma of mortality.”4 The Enlightenment is no exception; it is only a transition towards the destiny of the West. A new philosophy becomes necessary at the end of this transition. However, it is difficult to identify what this philosophy might be due to the rapid transformation of geopolitics in the twenty-first century, driven by several remarkable events such as 9/11, which revealed the vulnerability of the West, and the rise of China, which is silently reconfiguring the world order through its development plans in places like Africa, Latin America, and the South Pacific.
|
119 |
+
|
120 |
+
In his article, Kissinger is right when he says that Enlightenment philosophy was spread—or more precisely, universalized—by modern technology. However, he fails to mention that the Enlightenment was not simply an intellectual movement promoting reason and rationality, but also a fundamentally political movement.5 Navigational and military technology allowed European powers to colonize the world, leading to what we now call globalization. We have been taught that the Enlightenment as a whole aimed to fully realize humanity and universal values by fighting superstition (not necessarily religion), and that it was through science and technology that this battle was supposed to be won. In addition to creating new nautical and cartographic tools, the Enlightenment was also itself a process of orientation that situated the West as the center of this transformation, the source of its universalization.
|
121 |
+
|
122 |
+
Even as modern technology spreads Enlightenment thought, its own process of self-realization leads to self-negation: the dialectic of Enlightenment from a geopolitical point of view. In his short 1931 book Man and Technics, Oswald Spengler argued that the West was making a huge mistake by exporting its technology:
|
123 |
+
|
124 |
+
At the close of last century, the blind will-to-power began to make its decisive mistakes. Instead of keeping strictly to itself the technical knowledge that constituted their greatest asset, the “white” peoples complacently offered it to all the world, in every Hochschule, verbally and on paper, and the astonished homage of Indians and Japanese delighted them.6
|
125 |
+
|
126 |
+
As a result, continues Spengler, the Japanese became “technicians of the first rank, and in their [1904–5] war against Russia they revealed a technical superiority from which their teachers were able to learn many lessons.”7 Japan exposed the dilemma of technological globalization: on one hand, the spread of technology constructs a global axis of time though which European modernity becomes the synchronizing metric of all civilizations; on the other, the same spread frees modern science and technology from being the exclusive asset of European modernity, rendering the West vulnerable to global competition. As Hegel pointed out in The Phenomenology of Spirit, Enlightenment faith replaces religious faith without realizing itself to be also only a faith. In this way, Enlightenment thinking led us down a long road to globalization, all the while being defeated by its own negation. This would be a perfect postcolonial critique of the West; however, the story is not so simple.
|
127 |
+
|
128 |
+
2. The Constitution of the Global Axis of Time and its Apocalyptic End
|
129 |
+
Kissinger is wrong—the Enlightenment has not ended. Indeed, technology that is used for surveillance can also facilitate freedom of speech, and vice versa. However, let’s step outside of this anthropological and utilitarian reading of technology and take modern technology as constituting specific forms of knowledge and rationality.8 Modern technology—the support structure of Enlightenment philosophy—has become its own philosophy. Just as Marshall McLuhan stated that “the medium is the message,” so has the universalizing force of technology become the political project of the Enlightenment. As technology assumes and even performs the role of Enlightenment thinking, the medium ceases to be the carrier of meaning and instead becomes meaning itself—the knowledge through which progress is assured. After long celebrating democracy as an unshakable universal Western value, Donald Trump’s victory seems to have dissolved its hegemony into comedy. Suddenly, American democracy appears no different from bad populism. Especially when the leader of the Republican Party publicly declares his admiration for Kim Jong-un’s dictatorship, it can be tempting to agree with Kissinger that the period of Enlightenment thinking, along with the republicanism to which Kant aspired, has come to a close.
|
130 |
+
|
131 |
+
But we cannot be satisfied with such a naive caricature of the Enlightenment. It is unfair to claim that writers such as Voltaire insisted only on the superior value of the West without paying attention to cultural differences; for instance, Voltaire also praised the greatness of China’s four-thousand-year-old culture, as well as its emperor, who was an expert in astronomy.9 It seems astonishing, but Johann Gottfried von Herder turned such interest into a weapon against Voltaire himself, accusing the Frenchman of a lack of sensitivity to cultural differences, of being too eager to apply the classification and generalization of scientific methods to other cultures.10 However, it is true that cultural differences had fewer political implications for Voltaire than for Herder.
|
132 |
+
|
133 |
+
The universal values proposed by the philosophes, as Kissinger rightly points out, could be spread worldwide only through modern technology. At the same time, such technology has ended or accomplished the Enlightenment and is now going its own way, creating the need for a new guiding philosophy. What could this philosophy be? A transhumanist philosophy? An Eurasian conservative revolution? A Landian accelerationism, or maybe a leftist version of the same, both hoping to overcome capitalism by accelerating its contradictions until it self-destructs? Henri de Saint-Simon once believed that speeding up industrialization and improving transportation networks would make socialism possible, since resources and goods would be more evenly distributed.11 In his essay “Paris, Capital of the Nineteenth Century,” Walter Benjamin pointed out that Saint-Simon’s followers “had anticipated the development of the world-economy, but not the class-struggle.”12 A well-constructed rail network may actually reinforce inequality because it can distribute capitalist resources more efficiently. Acceleration in this sense is only a way to push Enlightenment universalism further. How can accelerating technology lead to the end of capitalism if it only creates another process of deterritorialization? One may argue for an absolute deterritorialization, but that would be Hegel approaching Deleuze from behind and giving him a monstrous child. Some claim that technology has surpassed capitalism itself, but this assumes that capitalism is a human-like creature who can be disrupted and rendered obsolete by technology, like an old man who doesn’t know how to send emails anymore after switching from a PC to a Mac.
|
134 |
+
|
135 |
+
As an aside, one must admit that technological acceleration is historically necessary for globalization, since non-Western countries have been able to enter the Western-dominated geopolitical arena only by creating a cost-effective assemblage of modern technology, cheap labor, and cheap nature. Both André Leroi-Gourhan and Gilbert Simondon pointed out that groups with advanced industrial technology were able to scale up their influence over groups with preindustrial technology.13 For Simondon especially, the revolt of minority groups against technology in the name of culture misunderstands the role of technology, since he sees a rationality in technology that transcends the limits of cultural difference. More importantly, Simondon holds out hope that the increasing perfection of technology will provide new perspectives for resolving the problem of alienation and the antagonism between culture and technology. However, the issue is far more complicated than Simondon’s optimism admits. In the colonization and modernization process, technological differences also maintain and reinforces power differences.
|
136 |
+
|
137 |
+
But what if the situation is reversed today, when, as Spengler said, the West is, or at least seems to be, surpassed by its students—a situation that will only continue? Take China as an example. Deng Xiaoping’s accelerationist politics has given China a leading role in the new millennium; Shenzhen has become China’s Silicon Valley and one of the craziest urban experiments in the world today. It is through technological acceleration and its accompanying economic triumph that we now witness the newest geopolitical arrangement to emerge since the Cold War: the East outstripping the West through digital innovation and automation. It is for this reason that Donald Trump has claimed that China has been stealing jobs from the US: jobs that were outsourced to China for its cheap labor are now being taken over by machines.
|
138 |
+
|
139 |
+
This technological acceleration is not a rupture, but the continuation of the Enlightenment. Kissinger’s article ignores the fact that technology, which embodies rationality and epistemology, is the true universal. That’s why he interprets the contemporary situation as the end of the Enlightenment, instead of its continuation in other forms. Are we then counterposing the universal and the relative? Or is this opposition itself the problem? It is beyond our scope here to deliver a treatise on the universal (though this remains an avoidable task). The desire to essentialize and establish the universal as a ground leads us to identify it as a substantial being, rather than seeing it as a dimension of existence. The relativists react by rejecting the universal without being able to integrate it into the particular. This oppositional thinking is at the core of both left and right populism. The same goes for the notion of humanity. By substantializing the human as a universal that transcends all particularities of culture and nature, we arrived at a humanism that is tantamount to nihilism. To get out of this impasse, we must first of all suspend the notion of humanity that has been handed down to us. And here we may want to invoke Carl Schmidt’s critique in The Concept of the Political: “The concept of humanity is an especially useful ideological instrument of imperialist expansion, and in its ethical-humanitarian form it is a specific vehicle of economic imperialism. Here one is reminded of a somewhat modified expression of Proudhon’s: whoever invokes humanity wants to cheat.”14
|
140 |
+
|
141 |
+
To reject the concept of humanity is to shatter the illusion created by a unifying discourse of the human, which belongs to a process of modernization qua synchronization. Modern technology synchronizes non-Western histories to a global time-axis of Western modernity. As both opportunity and problem, the synchronization process allows the world to enjoy science and technology, but it also draws the world into the global time-axis which, animated by humanism, is moving towards an apocalyptic end, whether it be the technological singularity, the “intelligence explosion,” or the emergence of “superintelligence.” Martin Heidegger already described this global time-axis in 1967: “The end of philosophy proves to be the triumph of the manipulable arrangement of a scientific-technological world and of the social order proper to this world. The end of philosophy means: the beginning of the world-civilization based upon Western European thinking.”15
|
142 |
+
|
143 |
+
Orientalists may respond with an uncanny smile: what an exaggerated statement! But the truth easily emerges when we observe the technical apparatus surrounding us and the gigantic force that is pushing us towards an apocalyptic end. What Heidegger calls the “end of philosophy” is nothing but the victory of the anthropological machine, the victory of a humanism that aspires to reinvent Homo sapiens as Homo deus through technological acceleration. Neoreactionaries and transhumanists celebrate artificial intelligence in the name of a posthumanist triumphalism, because superintelligence and technological singularity demonstrate the “possibility of sublime humanity.”
|
144 |
+
|
145 |
+
The so-called Dark Enlightenment is an effort to push the Heideggerian “end of philosophy” to the brink through a catastrophic intelligence explosion. Like in the I Ching, where the Pi hexagram (bad luck) is followed by the Tai hexagram (good luck) in a turn from extreme bad to good, this explosion will force the West to reinvent itself—or so its advocates believe. In their affirmation of a messianic acceleration towards the abyss, they conceive of themselves as anti-humanists. But what is behind this abyss? Robin Mackay has rightly pointed out that the fatal mistake of this vision of accelerationism “was to believe that, on the horizon of the deterritorialization opened up by capital, there would be disclosed an originary desire that could flow free of instituted structures of power.”16 One speculates on this unknown end of absolute deterritorialization like gamblers staring at casino tokens. Accelerating disorientation does not create an exit from the global time-axis. On the contrary, it merely disrupts momentarily the established orders and conventional modes of operation. In China, for example, expanding bandwidth and storage capacity for data flow have given rise to social credit systems, which simply stabilize and reterritorialize the flow of capital. A recent survey conducted by the Free University of Berlin showed that 80 percent of Chinese respondents approved or highly approved of these social credit systems, with 19 percent neutral and only 1 percent opposed.17 The disruptive and apocalyptic qualities intrinsic to acceleration are by no means anti-humanist. In fact, they reveal an extreme humanism fighting to save itself through massive destruction—a twenty-first-century nihilism.
|
146 |
+
|
147 |
+
Is it even possible to escape the synchronizing global time-axis of Western modernity, without proposing a deceleration, as sociologists such as Hartmut Rosa do? Are we capable of undoing its hold in order to advance its achievements in other directions?
|
148 |
+
|
149 |
+
We need to return to the word “acceleration” itself, since it is too easy to be fooled by an unexamined relation between acceleration and speed. If we recall high school physics, where a = v1-v2/t, acceleration is equal to the change of velocity (from v1 to v2) divided by time. V is velocity, not speed. Velocity is a vector having both magnitude and direction, while speed is mere magnitude. Why not consider another form of acceleration that does not push speed to its extreme, but rather changes the direction of movement, giving technology a new frame and orientation with regard to time and technological development? By so doing, we can also imagine a bifurcation of the future, which instead of moving towards the apocalypse, diverges from it and multiplies. But what does it mean to give technology a new frame? In order to do so, it is necessary to reflect on how we might reappropriate modern technology by systematically reflecting and working on the question of epistemologies and epistemes in light of multiple cosmotechnics, or simply put, the technodiversity that is historically traceable and still productive. This is a project I began with my book The Question Concerning Technology in China: An Essay on Cosmotechnics (2016), in which I use China as an example to elaborate on different conceptualizations of technology and the possibility of conceiving such a technodiversity in history and for the future. The proposal of multiple cosmotechnics—which is not, of course, limited to China—calls for us to reopen the concept of “technics” and reexamine the conditions of technical evolution.
|
150 |
+
|
151 |
+
|
152 |
+
3. Technodiversity and the Bifurcations of the Future
|
153 |
+
Technics is anthropologically universal in the process of hominization—the understanding of the human as a species because it is the exteriorization of memory and the liberation of human organs. With drawing and writing, human beings exteriorized their memories and imaginations; by inventing flint, the ancients liberated their fingers from many activities. We do not reject the notion that there is a universal dimension to technology, but it is only one of the dimensions. From a cosmotechnical standpoint, technics is fundamentally motivated and constrained by particular geographical and cosmological specificities. If we want to respond to the prospect of global self-extinction, we need to return to a carefully elaborated discourse on locality and the places of the human in the cosmos. In order to do so, we need first of all to reopen the question of technology, to conceive of multiple cosmotechnics instead of merely two: a premodern technics and a modern technics. To be sure, we must be careful with the word “locality” and its politics. Nostalgic invocations of tradition or culture can become problematic returns to nationalism, cultural essentialism, and ethnofuturism, when not approached dialectically. Here we are not considering small groups revolting against modern technologies in the name of culture or nature; rather, we are elaborating a general strategy to reappropriate technology by first of all affirming the irreducible multiplicity of technicity. While Simondon has been an inspiration for the concept of cosmotechnics, his own critique fails to articulate technics beyond the tradition of Western Enlightenment humanism he inherited.
|
154 |
+
|
155 |
+
To propose a pluralism is a gesture which could be attributed to both reactionaries and revolutionaries. Take the example of Herder, the fiercest opponent of Voltaire and the author of the book-length 1774 essay This Too a Philosophy of History for the Formation of Humanity, which argues that cultural experiences, values, and feelings are irreducibly diverse. Can one call Herder a nationalist? Many do consider him—a Lutheran priest, student of Kant, and mentor to Goethe—to be a founding figure of German nationalism and the Volksgeist. However, this view is not universally shared. Meineke once asked: “Did not Herder, when he arose to create a new epoch, proclaim both humanity and nationality?”18 Philosophers like Hans-Georg Gadamer and Isaiah Berlin also saw in Herder both a populism and pluralism, or as Charles Taylor put it, a populism and an “expressivism.”19 Herder is considered by some to be a genuine cosmopolitan thinker who roots cosmopolitanism in heterogeneity rather than homogeneity; he affirms differences not by claiming that each culture has a unique essence, but by arguing for the importance of locality and the equality of all cultures.
|
156 |
+
|
157 |
+
Humans are formed in distinct symbolic and linguistic worlds. Their different forms of knowledge and their different relations to the world and to the earth are not measurable by their advances in modern science and technology. The end of the Enlightenment has to begin by appropriating Herder after Gadamer, Berlin, and Taylor, since theirs is only the first step. We will have to understand the transformative power of heterogeneity instead of retreating to a certain Volk and depending on empathy or sensitivity to resolve tensions within increasingly isolated groupings. As a response to the ecological problems associated with the Anthropocene, anthropologists such as Philippe Descola and others have reopened the question of radical pluralism in a way that considers what is called “multinaturalism” instead of multiculturalism. Because naturalism, which counterposes nature and culture, is very much a product of modernity, it does not capture how nonhumans are perceived in other parts of the world. However, with modernization as a synchronization process, we encounter a tipping point that reopens concepts such as nature and technics which have been inherited as universal without being questioned. This call for pluralism is for us a reminder to consciously reappropriate modern science and technology, to give it a new direction at a time when its planetary spread opens up such a possibility.20
|
158 |
+
|
159 |
+
On the other hand, we may understand Kissinger’s end-of-Enlightenment claim as marking the full realization of a single global axis of time in which all historical times converge into the synchronizing metric of European modernity. It is the moment of disorientation—a loss of direction as well as of the Orient in relation to the Occident. The unhappy consciousness of fascism and xenophobia arises from this inability to orient: as a response, it offers an easy identity politics and an aestheticized politics of technology.
|
160 |
+
|
161 |
+
More broadly, such a disorientation can be seen as a desirable and necessary deterritorialization of contemporary capitalism, which facilitates accumulation beyond temporal and spatial constraints. War is the technique of disruption par excellence, vastly more effective than Uber and Airbnb. In his 1933 The Hour of Decision: Germany and World-Historical Evolution, Spengler describes the war machine as the only possible response to the geopolitical crisis of the time: “England gained her wealth by battles and not by bookkeeping and speculation … [Germany] had to carry on its wars with foreign money and in the service of that money, and it waged wars over miserable scraps of its own country that one diminutive state took from another.”21
|
162 |
+
|
163 |
+
The prospect of war as a solution was not pursued only in the West: the Kyoto school philosophers also proposed total war as a means of overcoming modernity.22 Today, could global competition over the development of artificial intelligence and space technology become the new condition of such a war? As Spengler wrote in 1933, certain forces are dragging us backwards. It is worth noting the major similarities between his epoch and ours, but we also need to pay special attention to the differences. Spengler wrote in The Hour of Decision of a certain dogmatic thinking in non-Western civilizations that emerged with modernity and was associated with a colonial mentality:
|
164 |
+
|
165 |
+
Immemorially old “Fellaheen” peoples such as the Indian and Chinese can never again play an independent part in the world of the great powers. They can change their masters, drive one out—as, for example, the Englishman from India—but it is only to succumb to another. They will never again produce a form of political existence of their own. For that they are too old, too rigid, too used up.23
|
166 |
+
|
167 |
+
This failure is largely due to the fact that the question of technology has never been sufficiently addressed, neither in the West nor elsewhere: technology remains a utility, and there is no way of seeing the kingdom of ends beyond the limits of utility and efficiency. Efficiency is a very important factor of technological innovation, but it has to be measured according to a long-term vision instead of short-term profits. The other thing that holds back the colonial mentality is a cynicism that sees no way out. After all, who can escape the economic and geopolitical competition to master artificial intelligence when technological linearity is identified with the progress of humanity? We can be certain that artificial intelligence will have a significant impact on our societies and economies. If China or Russia slowed their pace of technological innovation, they would lose their competitive edge: Putin already declared to a room full of Russian school children on September, 1 2017 that “whoever leads in AI will dominate the world.”24 But if technological acceleration and innovation are the common task of the sovereign and capital, human cynicism will only deepen as we feel increasingly helpless in the face of technological systems that displace the human roles in so many processes. True philosophical thinking can be the only response to this aporia.
|
168 |
+
|
169 |
+
I don’t mean to suggest that modern science and technology are evil (not least because they were my first areas of study). Nor am I suggesting that non-European cultures and traditions have been destroyed by evil modern technologies imposed by the West, and that therefore we should give up modern science and technology. The question, rather, is how this historical process can be rethought, and what futures are still available for imagination and realization. If we identify Enlightenment thought with modern technology as an irreversible process guided by universality and rationality, then the only question that remains to be asked is: To be or not to be? But if we affirm that multiple cosmotechnics exist, and that these may allow us to transcend the limit of sheer rationality, then we can find a way out of never-ending modernity and the disasters that have accompanied it. It would be tragic to misunderstand rationality merely as strict and rigid reasoning—unfortunately, it has been often mistaken as such. The history of reason and its relation to nature and technology, from Leibniz to cybernetics and machine learning, has to be constructed and interrogated differently than it has been.25
|
170 |
+
|
171 |
+
Certain reflections on culture may provide a way to understand these different modes of technological thinking. To rediscover multiple cosmotechnics is not to refuse artificial intelligence or machine learning, but to reappropriate modern technology, to give other frames to the enframing (Gestell) at the core of modern technology.26 If we want to surpass modernity, there is no way to simply reset it as if it were a computer or a smartphone. We must instead escape its global time-axis, escape a (trans)humanism that subordinates other beings to the terms of its own destiny, and propose a new agenda and imagination of technology that open up new forms of social, political, and aesthetic life and new relations with nonhumans, the earth, and the cosmos. All of this remains to be thought, since it demands a Nietzschean revaluation of the question of technology, and this is possible only when done collectively.
|
172 |
+
|
173 |
+
In this sense, we can take Kissinger’s statement not as a target of criticism, but as an invitation to think beyond the end of the Enlightenment, as a challenge to undertake the task of thinking through the plurality of its forms. Perhaps Kissinger’s own closing warning is the most appropriate way to end this critique of him: “If we do not start this effort soon, before long we shall discover that we started too late.”
|