NOT WITHOUT US

A Challenge to Computer Professionals to Help Bring the Present

Insanity to a Halt

By Joseph Weizenbaum, professor of computer science at MIT

(This is an English translation of a talk given in German to the

Association of Computer Professionals in West Germany in July 1986.

You are welcome to reproduce and distribute it.)

 

Whenever I come to Europe, especially to West Germany, I am

amazed by the normality of everyday life: superhighways, "music" that

assaults one in restaurants, the many parks, the forests of television

antennas on the roofs of houses and so on. I am amazed because of

Europe's geographic position and all that follows from it. In West

Germany, for example, there is the border with the other Germany,

dense with military installations of all sorts. There are holes in

the street that are intended to be filled with nuclear land mines if

Russian tanks should come. These are signs of Europe's physical and

psychological proximity to the final catastrophe.

We in America are, in a certain sense, no more distant from the

catastrophe than the Europeans are. Not only Chernobyl, but also the

threat of war is everywhere. And war is everyone's enemy. In case of

war, regardless of whether unintentionally initiated by technology

allegedly designed to avert war, or by so-called statesmen or women

who thought it their duty to push the button, you may die ten minutes

earlier than we in fortress America, but we shall all die.

But we have no holes in our streets for atomic land mines that

are intended to delay Soviet tank regiments. We see our missile silos

only now and then -- that is, only whenever it pleases someone to show

them to us on television. No matter how passionately our government

tries to convince us that the nasty Soviets are effectively as near to

us as to Europeans, that they threaten us from Cuba and Nicaragua,

Americans are, on the whole, quite unconvinced and untroubled by such

efforts. The American experience of war has allowed us to develop an

"it can't happen here" attitude, rather than a concrete fear of what

appears to be far removed from the immediate concerns of daily life.

We know that it is emotionally impossible for anyone to live for

very long in the face of immediate threats to existence without

bringing to bear psychological mechanisms that will exclude these

dangers from consciousness, permitting them to surface only rarely.

But when repression necessitates systematically misdirected efforts,

or excludes potentially life-saving behavior, then it is time to

replace it with a conscious effort to find the prod to correct action.

That time has come for computer professionals. We now have the

power radically to turn the state of the world in directions conducive

to life.

In order to gain the necessary courage (not all of us are saints

or heroes) we have to understand that for us as individuals, as well

as for those we love, our present behavior is far more dangerous, even

life threatening, than what healthy common sense now demands of us.

None of the weapons that today threaten every human being with murder,

and whose design, manufacture and sale condemns countless people to

starvation, could be developed without the earnest cooperation of

computer professionals. Without us, the arms race, especially the

qualitative arms race, cannot march another step.

What does this say to us?

First, that we computer experts -- as well as specialists in many

other technical domains -- share in the guilt of having brought about

the present dangerous state of the world. Those among us who, perhaps

without being aware of it, devote our talents and strengths to death

rather than to life have little right to curse politicians, statesmen

and women for not bringing us peace. It isn't enough to make pretty

posters that can be carried in demonstrations. Those who carry them

must care whether their daily work helps to make possible the very

devices the use of which they are protesting.

At this point, the domain called Artificial Intelligence (AI)

comes especially to mind. Many of the technical tasks and problems in

this subdiscipline of computer science stimulate the imagination and

creativity of technically oriented workers particularly strongly.

Goals like making a thinking being out of the computer, giving the

computer the ability to understand spoken language, making it possible

for the computer to see, offer nearly irresistible temptations to

those among us who have not fully sublimated our playful sandbox

fantasies, or who mean to satisfy our delusions of omnipotence on the

computer stage. Such tasks are extraordinarily demanding and

interesting. Robert Oppenheimer called them sweet. Besides, research

projects in these areas are generously funded. The required moneys

usually come out of the coffers of the military, at least in America.

It is enormously tempting and, in Artificial Intelligence work,

seductively simple to lose or hide oneself in details, in subproblems

and their subproblems and so on. The actual problems on which one

works -- and which are so generously supported -- are disguised and

transformed until their representations are mere fables: harmless,

innocent, lovely fairy tales.

Here is an example. A doctoral student characterized his

projected dissertation task as follows. A child, six or seven years

old, sits in front of a computer display that shows a kitten and a

bear, in full color. The kitten is playing with a ball. The child

speaks to the computer system: "The bear should say 'thank you' when

someone gives him something." The system responds in a synthetic, but

nevertheless pleasing voice: "Thank you, I understand." Then the

child again: "Kitty, give your ball to your friend." Immediately we

see the kitten on the computer display throw the ball to the bear.

Then we hear the bear say: "Thank you, my dear kitten."

This is the kernel of what the system, development of which is to

constitute the student's doctoral work, is to accomplish. Seen from a

technical point of view, the system is to understand spoken

instructions -- that alone is not simple -- and translate them into a

computer program which it is then to integrate seamlessly into its own

computational structure. Not at all trivial, and beyond that, quite

touching.

Now a translation to reality. A fighter pilot is addressed by

his pilot's assistant system: "Sir, I see an enemy tank column below.

Your orders, please." The pilot: "When you see something like that,

don't bother me, destroy the bastards and record the action. That's

all." The system answers: "Yes, sir!" and the plane's rockets fly

earthward.

This pilot's assistant system is one of three weapons systems

that are expressly described, mainly as a problem for artificial

intelligence, in the Strategic Computing Initiative, a new major

research and development program of the American military. Over

$600,000,000 are to be spent on this program in the next four or five

years.

It isn't my intention to assail or revile military systems at

this point. I intend this example from the actual practice of

academic artificial intelligence research in America to illustrate the

euphemistic linguistic dissimulation whose effect it is to hinder

thought and, ultimately, to still conscience.

I don't know whether it is especially computer science or its

subdiscipline Artificial Intelligence that has such an enormous

affection for euphemism. We speak so readily of computer systems that

understand, that see, decide, make judgments, and so on, without

ourselves recognizing our own superficiality and immeasurable naivete

with respect to these concepts. We anesthetize our ability to

evaluate the quality of our work and, what is more important, to

identify and become conscious of its end use.

The student mentioned above imagines his work to be about

computer games for children, involving perhaps toy kittens, bears and

balls. Its actual and intended end use will probably mean that some

day a young man, quite like the student himself -- someone with

parents and possibly a girl friend -- will be set afire by an

exploding missile sent his way by a system shaped by his own research.

The psychological distance between the student's conception of his

work and its actual implications is astronomic. It is precisely that

enormous distance that makes it possible not to know and not to ask if

one is doing sensible work or contributing to the greater efficiency

of murderous devices.

One can't escape this state without asking, again and again:

"What do I actually do? What is the final application and use of my

work? Am I content or ashamed to have contributed to this use?"

I am reminded in this context of a well known American journalist

who, during a Middle East highjacking, suggested that under certain

circumstances the Israelis shoot ten Arab prisoners and, should the

circumstances not change, shoot ten more the next day, and so on. He

should not have made this suggestion unless he was prepared to go

personally among the prisoners and look into the eyes of the men, some

of whom would hear him say, "You, you will die today." He should have

been prepared as well to hold the pistol to the heads of those he

selected and command his own finger to pull the trigger.

Just so should we ask ourselves about our own work. Once we have

abandoned the prettifying of our language, we can begin to speak among

ourselves realistically and in earnest about our work as computer

professionals.

"You, colleague of many years, you are working on a machine

consisting of two to the fifteenth and more microprocessors running

simultaneously. With the help of such a machine one can first

simulate then construct much more efficient, smaller and lighter

hydrogen bombs. Imagine, for a moment, you were an eyewitness at

Hiroshima in 1945; you saw people stripped of their skin die. Would

you want to make this happen thousands of times more? Would you so

torture a single human being with your own hands? If you would not,

regardless of what end would be served, then you must stop your work."

One should ask similar questions with respect to other branches

of computer science, for example, with respect to attempts to make it

possible for computer systems to see. Progress in this domain will be

used to steer missiles like the Cruise and Pershing ever more

precisely to their targets, where murder will be committed.

Many will argue that the computer is merely a tool. As such it

can be used for good or evil. In and of itself, it is value free.

Scientists and technicians cannot know how the products of their work

will be applied, whether they will find a good or an evil use. Hence

scientists and technicians cannot be held responsible for the final

application of their work.

That point of view is manifested in the world famous Draper

Laboratory, next door to the MIT building where I work. Draper is

devoted almost entirely to missile guidance and submarine navigation.

Many of the scientists employed there argue that the systems they work

on can take men to the moon and bring them back, as well as guarantee

that missiles aimed at Moscow will actually hit Moscow, their target.

They cannot know in advance, they say, which of these two or still

other goals their work will serve in the end. How then can they be

held responsible for all the possible consequences of their work?

So it is, on the whole, with computer professionals. The

doctoral student I mentioned, who wishes to be able to converse with

his computer display, does in fact believe that future applications of

his work will be exclusively in innocent applications like children's

games. Perhaps his research is not sponsored by the Pentagon's

Strategic Computing Initiative; perhaps he never even heard of SCI.

How then can he be held responsible if his work is put to anti-human

use?

Here is where we come to the essence of the matter. Today we

know with virtual certainty that every scientific and technical result

will, if at all possible, be put to use in military systems.

The computer, together with the history of its development, is

perhaps the key example. But we should also think in this connection

of everything that has to do with flight, or of things atomic, of

communications systems, satellites, space ships, and most of the

scientific achievements of the human genius. We may then convince

ourselves that in the concrete world in which we live, the burden of

proof rests with those who assert that a specific new development is

immune from the greed of the military.

In these circumstances, scientific and technical workers cannot

escape their responsibility to inquire about the end use of their

work. They must then decide, once they know to what end it will be

used, whether or not they would serve these ends with their own hands.

I don't believe the military, in and of itself, to be an evil.

Nor would I assert that the fact that a specific technology that has

been adopted by the military is, on that ground alone, an evil. In

the present state of the evolution of the sovereign nation-state -- in

other words, in the insane asylum in which we live -- each state needs

a military just as every city needs a fire department. But no one

pleads for a fire station on every corner, and no one wishes for a

city fire department that makes a side business of committing arson in

the villages adjacent to the city.

But we see our entire world, particularly its universities and

science and engineering facilities, being more profoundly militarized

every day. "Little" wars burn in almost every part of the earth.

(They serve, in part, to test the high tech weapons of the "more

advanced nations.") More than half of all the earth's scientists and

engineers work more or less directly in military institutions, or in

institutions supported by the military. That is an evil that must be

resisted.

We must also recognize that it is only our already internalized

habit of prettifying our language, in order not to arouse our

conscience, that permits us to speak in terms of weapons and weapons

delivery systems at all, when we are, in fact, discussing atomic

explosives and hydrogen bombs. Those aren't weapons, they are mass

murder machines and mass murder machine delivery systems. That is how

we should speak of them: clearly, distinctly, and without evasion.

Once we recognize that a nuclear mass murder machine is nothing other

than an instant Auschwitz -- without railroads or Eichmanns or Dr.

Mengele, but an Auschwitz just the same -- can we continue then to

work on systems that steer these devices to living cities?

That is the question I ask. Each of us must earnestly ask

ourselves such questions and deeply consider the responses we find in

ourselves. Our answers must finally manifest themselves in our

actions -- concretely, in what we do every day.

Probably the most pandemic mental illness of our time is the

almost universally held belief that the individual is powerless. This

self-fulfilling delusion will surely be offered as a counter argument

to my theses. I demand, do I not, that a whole profession refuse to

participate in the murderous insanity of our time. "That cannot be

effective," I can already hear it said," That is plainly impossible.

After all, if I don't do it, someone else will." First, and on the

most elementary level, "If I don't do it, someone else will" cannot

serve as a basis of moral behavior. Every crime imaginable can be

justified with those words. For example: If I don't steal the

sleeping drunk's money, someone else will. But it is not at all

trivial to ask after the meaning of effectiveness in the present

context. Surely, effectiveness is not a binary matter, an either/or

matter. To be sure, if what I say here were to induce a strike on the

part of all scientists with respect to weapons work, that would have

to be counted as effective. But there are many much more modest

measures of effectiveness.

I think it was George Orwell who once wrote, "The highest duty of

intellectuals in these times is to speak the simplest truths in the

simplest possible words." For me that means, first of all, to

articulate the absurdity of our work in my actions, my writings and

with my voice. I hope thereby to stir my students, my colleagues,

everyone to whom I can speak directly. I hope to encourage those who

have already begun to think similarly, and to be encouraged by them,

and possibly rouse others out of their slumber. Courage, like fear is

catching.

Even the most modest success in such attempts has to be counted

as effective. Beyond that, in speaking as I do, I put what I discuss

here on the public agenda and contribute to its legitimation. These

are modest goals that can surely be reached.

But, finally, I want to address such larger goals as, for

example:

Ridding the world of nuclear mass murder devices and perhaps also

of nuclear power generators.

So reordering the world that it becomes impossible ever again to

convince workers of one country that it is a necessity of life that

they feed their families on the flesh and the blood and the tears of

people of other countries. (That is, unfortunately, the fate of many

workers today, and not only those who earn their daily bread in

armaments factories, but equally those of us whose daily work is to

sharpen high tech weapons.)

So reordering the world that every human being has available to

him or herself all material goods necessary for living in dignity. (I

have often heard well-meaning people say that, if we apply technology,

especially computer and communications technology wisely, we may reach

this goal in perhaps 50 to 100 years. But we can reach it sooner, and

without waiting for technological advances. For the obstacle is not

the absence of technology, it is the absence of political will.)

I once heard Elie Wiesel say: "We must believe the impossible is

possible." I understood that in two different ways. First, had we

been able to believe that "the land of the poets and the thinkers"

could give birth to human extermination factories, we might not have

had to experience Bergen Belsen. The impossible horror proved

possible and became reality.

But there is a more hopeful interpretation. It seemed impossible

in the America of only 150 years ago ever to abolish the slavery of

the black people. The entire economy of America's south was built on

cotton. Cotton could neither be planted nor harvested, it was

believed, without the unpaid toil of thousands of human beings out of

whose wretchedness the plantation master could squeeze his profit.

Nevertheless, at first only a few farseeing men and women, dreamers

all, in Massachusetts, later many more citizens, came to believe the

impossible was possible, that the slaves could be freed and slavery

ended.

The impossible goals I mention here are possible, just as it is

possible that we will destroy the human race. I alone can neither

achieve the one nor prevent the other. But neither can it be done

without me, without us.

I have no right to demand anything from my colleagues. But they

must know that we have the power either to increase the efficiency of

the mass murder instruments we have and thereby make the murder of our

children more likely, or to bring the present insanity to a halt, so

that we and our children have a chance to live in human dignity.

Let us think about what we actually accomplish in our work, about

how it will be used, and whether we are in the service of life or

death.

Computers for Christ - Chicago


Index of Articles of Interest    Home

  1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 | 10 | 11 | 12 | 13 | 14 | 15 | 16 | 17 | 18 | 19 | 20 | 21 | 22 | 23 | 24 | 25 | 26 | 27 | 28 | 29 | 30 | 31 | 32 | 33 | 34 | 35 | 36 | 37 | 38 | 39 | 40 | 41 | 42 | 43 | 44 | 45 | 46 | 47 | 48 | 49 | 50 | 51 | 52 | 53 | 54 | 55 | 56 | 57 | 58 | 59 | 60 | 61 | 62 | 63 | 64 | 65 | 66 | 67 | 68 | 69 | 70 | 71 | 72 | 73 | 74 | 75 | 76 | 77 | 78 | 79 | 80 | 81 | 82 | 83 | 84 | 85 | 86 | 87 | 88 | 89 | 90 | 91 | 92 | 93 | 94 | 95 | 96 | 97 | 98 | 99 | 100 | 101 | 102 | 103 | 104 | 105 | 106 | 107 | 108 | 109 | 110 | 111 | 112 | 113 | 114 | 115 | 116 | 117 | 118 | 119 | 120 | 121 | 122 | 123 | 124 | 125 | 126 | 127 | 128 | 129 | 130 | 131 | 132 | 133 | 135 | 136 | 137 | 138 | 139 | 140 | 141 | 142 | 143 | 144 | 145 | 146 | 147 | 148 | 149 | 150 | 151 | 152 | 153 | 154 | 155 | 156 | 157 | 158 | 159 | 160 | 161 | 162 | 163 | 164 | 165 | 166 | 167 | 168 | 169 | 170 | 171 | 172 | 173 | 174 | 175 | 176 | 177 | 178 | 179 | 180 | 181 | 182 | 183 | 184 | 185 | 186 | 187 | 188 | 189 | 190 | 191 | 192 | 193 | 194 | 195 | 196 | 197 | 198 | 199 | 200 | 201 | 202 | 203 | 204 | 205 | 206 | 207 | 208 | 209 | 210 | 211 | 212 | 213 | 214 | 215 | 217 | 218 | 219 | 220 | 221 | 222 | 223 | 224 | 225 | 226 | 227 | 228 | 229 | 230 | 231 | 232 | 233 | 235 | 236 | 237 | 238 | 239 | 240 | 241 | 242 | 243 | 244 | 245 | 246 | 247 | 248 | 249 | 250 | 251 | 252 | 253 | 254 | 255 | 256 | 257 | 258 | 259 | 260 | 261 | 262 | 263 | 264 | 265 | 266 | 267 | 268 | 269 | 270 | 271 | 272 | 273 | 274 | 275 | 276 | 277 | 278 | 279 | 280 | 281 | 282 | 283 | 284 | 285 | 286 | 287 | 288 | 289 | 290 | 291 | 292 | 293 | 294 | 295 | 296 | 297 | 298 | 299 | 300 | 301 | 302 | 303 | 304 | 305 | 306 | 307 | 308 | 309 | 310 | 311 | 312 | 313 | 314 | 315 | 317 | 318 | 319 | 320 | 321 | 322 | 323 | 324 | 325 | 326 | 327 | 328 | 329 | 330 | 331 | 332 | 333

 

Home | Bible versions | Bible Dictionary | Christian Classics | Christian Articles | Daily Devotions

Sister Projects: Wikichristian | WikiMD

BelieversCafe is a large collection of christian articles with over 40,000 pages


Our sponsors:   W8MD sleep and weight loss center