Продолжая использовать наш сайт, вы даете согласие на обработку файлов cookie, которые обеспечивают правильную работу сайта. Благодаря им мы улучшаем сайт!
Принять и закрыть

Читать, слущать книги онлайн бесплатно!

Электронная Литература.

Бесплатная онлайн библиотека.

Читать: - на бесплатной онлайн библиотеке Э-Лит


Помоги проекту - поделись книгой:

Bruce Sterling

Essays. FSF Columns

OUTER CYBERSPACE

Dreaming of space-flight, and predicting its future, have

always been favorite pastimes of science fiction. In my first science

column for F&SF, I can't resist the urge to contribute a bit to this

grand tradition.

A science-fiction writer in 1991 has a profound advantage over

the genre's pioneers. Nowadays, space-exploration has a past as

well as a future. "The conquest of space" can be judged today, not

just by dreams, but by a real-life track record.

Some people sincerely believe that humanity's destiny lies in the

stars, and that humankind evolved from the primordial slime in order

to people the galaxy. These are interesting notions: mystical and

powerful ideas with an almost religious appeal. They also smack a

little of Marxist historical determinism, which is one reason why the

Soviets found them particularly attractive.

Americans can appreciate mystical blue-sky rhetoric as well as

anybody, but the philosophical glamor of "storming the cosmos"

wasn't enough to motivate an American space program all by itself.

Instead, the Space Race was a creation of the Cold War -- its course

was firmly set in the late '50s and early '60s. Americans went into

space *because* the Soviets had gone into space, and because the

Soviets were using Sputnik and Yuri Gagarin to make a case that

their way of life was superior to capitalism.

The Space Race was a symbolic tournament for the newfangled

intercontinental rockets whose primary purpose (up to that point) had

been as instruments of war. The Space Race was the harmless,

symbolic, touch-football version of World War III. For this reason

alone: that it did no harm, and helped avert a worse clash -- in my

opinion, the Space Race was worth every cent. But the fact that it was

a political competition had certain strange implications.

Because of this political aspect, NASA's primary product was

never actual "space exploration." Instead, NASA produced public-

relations spectaculars. The Apollo project was the premiere example.

The astonishing feat of landing men on the moon was a tremendous

public-relations achievement, and it pretty much crushed the Soviet

opposition, at least as far as "space-racing" went.

On the other hand, like most "spectaculars," Apollo delivered

rather little in the way of permanent achievement. There was flag-

waving, speeches, and plaque-laying; a lot of wonderful TV coverage;

and then the works went into mothballs. We no longer have the

capacity to fly human beings to the moon. No one else seems

particularly interested in repeating this feat, either; even though the

Europeans, Indians, Chinese and Japanese all have their own space

programs today. (Even the Arabs, Canadians, Australians and

Indonesians have their own satellites now.)

In 1991, NASA remains firmly in the grip of the "Apollo

Paradigm." The assumption was (and is) that only large, spectacular

missions with human crews aboard can secure political support for

NASA, and deliver the necessary funding to support its eleven-billion-

dollar-a-year bureaucracy. "No Buck Rogers, no bucks."

The march of science -- the urge to actually find things out

about our solar system and our universe -- has never been the driving

force for NASA. NASA has been a very political animal; the space-

science community has fed on its scraps.

Unfortunately for NASA, a few historical home truths are

catching up with the high-tech white-knights.

First and foremost, the Space Race is over. There is no more

need for this particular tournament in 1992, because the Soviet

opposition is in abject ruins. The Americans won the Cold War. In

1992, everyone in the world knows this. And yet NASA is still running

space-race victory laps.

What's worse, the Space Shuttle, one of which blew up in 1986,

is clearly a white elephant. The Shuttle is overly complex, over-

designed, the creature of bureaucratic decision-making which tried to

provide all things for all constituents, and ended-up with an

unworkable monster. The Shuttle was grotesquely over-promoted,

and it will never fulfill the outrageous promises made for it in the '70s.

It's not and never will be a "space truck." It's rather more like a Ming

vase.

Space Station Freedom has very similar difficulties. It costs far

too much, and is destroying other and more useful possibilities for

space activity. Since the Shuttle takes up half NASA's current budget,

the Shuttle and the Space Station together will devour most *all* of

NASA's budget for *years to come* -- barring unlikely large-scale

increases in funding.

Even as a political stage-show, the Space Station is a bad bet,

because the Space Station cannot capture the public imagination.

Very few people are honestly excited about this prospect. The Soviets

*already have* a space station. They've had a space station for years

now. Nobody cares about it. It never gets headlines. It inspires not

awe but tepid public indifference. Rumor has it that the Soviets (or

rather, the *former* Soviets) are willing to sell their "Space Station

Peace" to any bidder for eight hundred million dollars, about one

fortieth of what "Space Station Freedom" will cost -- and nobody can

be bothered to buy it!

Manned space exploration itself has been oversold. Space-

flight is simply not like other forms of "exploring." "Exploring"

generally implies that you're going to venture out someplace, and

tangle hand-to-hand with wonderful stuff you know nothing about.

Manned space flight, on the other hand, is one of the most closely

regimented of human activities. Most everything that is to happen on

a manned space flight is already known far in advance. (Anything not

predicted, not carefully calculated beforehand, is very likely to be a

lethal catastrophe.)

Reading the personal accounts of astronauts does not reveal

much in the way of "adventure" as that idea has been generally

understood. On the contrary, the historical and personal record

reveals that astronauts are highly trained technicians whose primary

motivation is not to "boldly go where no one has gone before," but

rather to do *exactly what is necessary* and above all *not to mess up

the hardware.*

Astronauts are not like Lewis and Clark. Astronauts are the

tiny peak of a vast human pyramid of earth-bound technicians and

mission micro-managers. They are kept on a very tight

(*necessarily* tight) electronic leash by Ground Control. And they

are separated from the environments they explore by a thick chrysalis

of space-suits and space vehicles. They don't tackle the challenges of

alien environments, hand-to-hand -- instead, they mostly tackle the

challenges of their own complex and expensive life-support

machinery.

The years of manned space-flight have provided us with the

interesting discovery that life in free-fall is not very good for people.

People in free-fall lose calcium from their bones -- about half a percent

of it per month. Having calcium leach out of one's bones is the same

grim phenomenon that causes osteoporosis in the elderly --

"dowager's hump." It makes one's bones brittle. No one knows quite

how bad this syndrome can get, since no one has been in orbit much

longer than a year; but after a year, the loss of calcium shows no

particular sign of slowing down. The human heart shrinks in free-

fall, along with a general loss of muscle tone and muscle mass. This

loss of muscle, over a period of months in orbit, causes astronauts and

cosmonauts to feel generally run-down and feeble.

There are other syndromes as well. Lack of gravity causes

blood to pool in the head and upper chest, producing the pumpkin-

faced look familiar from Shuttle videos. Eventually, the body reacts

to this congestion by reducing the volume of blood. The long-term

effects of this are poorly understood. About this time, red blood cell

production falls off in the bone marrow. Those red blood cells which

are produced in free-fall tend to be interestingly malformed.

And then, of course, there's the radiation hazard. No one in

space has been severely nuked yet, but if a solar flare caught a crew in

deep space, the results could be lethal.

These are not insurmountable medical challenges, but they

*are* real problems in real-life space experience. Actually, it's rather

surprising that an organism that evolved for billions of years in

gravity can survive *at all* in free-fall. It's a tribute to human

strength and plasticity that we can survive and thrive for quite a

while without any gravity. However, we now know what it would be

like to settle in space for long periods. It's neither easy nor pleasant.

And yet, NASA is still committed to putting people in space.

They're not quite sure why people should go there, nor what people

will do in space once they're there, but they are bound and determined

to do this despite all obstacles.

If there were big money to be made from settling people in

space, that would be a different prospect. A commercial career in

free-fall would probably be safer, happier, and more rewarding than,

say, bomb-disposal, or test-pilot work, or maybe even coal-mining.

But the only real moneymaker in space commerce (to date, at least) is

the communications satellite industry. The comsat industry wants

nothing to do with people in orbit.

Consider this: it costs $200 million to make one shuttle flight.

For $200 million you can start your own communications satellite

business, just like GE, AT&T, GTE and Hughes Aircraft. You can join

the global Intelsat consortium and make a hefty 14% regulated profit

in the telecommunications business, year after year. You can do quite

well by "space commerce," thank you very much, and thousands of

people thrive today by commercializing space. But the Space Shuttle,

with humans aboard, costs $30 million a day! There's nothing you can

make or do on the Shuttle that will remotely repay that investment.

After years of Shuttle flights, there is still not one single serious

commercial industry anywhere whose business it is to rent workspace

or make products or services on the Shuttle.

The era of manned spectaculars is visibly dying by inches. It's

interesting to note that a quarter of the top and middle management

of NASA, the heroes of Apollo and its stalwarts of tradition, are

currently eligible for retirement. By the turn of the century, more than

three-quarters of the old guard will be gone.

This grim and rather cynical recital may seem a dismal prospect

for space enthusiasts, but the situation's not actually all that dismal at

all. In the meantime, unmanned space development has quietly

continued apace. It's a little known fact that America's *military*

space budget today is *twice the size* of NASA's entire budget! This

is the poorly publicized, hush-hush, national security budget for

militarily vital technologies like America's "national technical means

of verification," i.e. spy satellites. And then there are military

navigational aids like Navstar, a relatively obscure but very

impressive national asset. The much-promoted Strategic Defence

Initiative is a Cold War boondoggle, and SDI is almost surely not long

for this world, in either budgets or rhetoric -- but both Navstar and

spy satellites have very promising futures, in and/or out of the

military. They promise and deliver solid and useful achievements,

and are in no danger of being abandoned.

And communications satellites have come a very long way since

Telstar; the Intelsat 6 model, for instance, can carry thirty thousand

simultaneous phone calls plus three channels of cable television.

There is enormous room for technical improvement in comsat

technologies; they have a well-established market, much pent-up

demand, and are likely to improve drastically in the future. (The

satellite launch business is no longer a superpower monopoly; comsats

are being launched by Chinese and Europeans. Newly independent

Kazakhstan, home of the Soviet launching facilities at Baikonur, is

anxious to enter the business.)

Weather satellites have proven vital to public safety and

commercial prosperity. NASA or no NASA, money will be found to

keep weather satellites in orbit and improve them technically -- not

for reasons of national prestige or flag-waving status, but because it

makes a lot of common sense and it really pays.

But a look at the budget decisions for 1992 shows that the

Apollo Paradigm still rules at NASA. NASA is still utterly determined

to put human beings in space, and actual space science gravely suffers

for this decision. Planetary exploration, life science missions, and

astronomical surveys (all unmanned) have been cancelled, or

curtailed, or delayed in the1992 budget. All this, in the hope of

continuing the big-ticket manned 50-billion-dollar Space Shuttle, and

of building the manned 30-billion-dollar Space Station Freedom.

The dire list of NASA's sacrifices for 1992 includes an asteroid

probe; an advanced x-ray astronomy facility; a space infrared

telescope; and an orbital unmanned solar laboratory. We would have

learned a very great deal from these projects (assuming that they

would have actually worked). The Shuttle and the Station, in stark

contrast, will show us very little that we haven't already seen.

There is nothing inevitable about these decisions, about this

strategy. With imagination, with a change of emphasis, the

exploration of space could take a very different course.

In 1951, when writing his seminal non-fiction work THE

EXPLORATION OF SPACE, Arthur C. Clarke created a fine

imaginative scenario of unmanned spaceflight.

"Let us imagine that such a vehicle is circling Mars," Clarke

speculated. "Under the guidance of a tiny yet extremely complex

electronic brain, the missile is now surveying the planet at close

quarters. A camera is photographing the landscape below, and the

resulting pictures are being transmitted to the distant Earth along a

narrow radio beam. It is unlikely that true television will be possible,

with an apparatus as small as this, over such ranges. The best that

could be expected is that still pictures could be transmitted at intervals

of a few minutes, which would be quite adequate for most purposes."

This is probably as close as a science fiction writer can come to

true prescience. It's astonishingly close to the true-life facts of the

early Mars probes. Mr. Clarke well understood the principles and

possibilities of interplanetary rocketry, but like the rest of mankind in

1951, he somewhat underestimated the long-term potentials of that

"tiny but extremely complex electronic brain" -- as well as that of

"true television." In the 1990s, the technologies of rocketry have

effectively stalled; but the technologies of "electronic brains" and

electronic media are exploding exponentially.

Advances in computers and communications now make it

possible to speculate on the future of "space exploration" along

entirely novel lines. Let us now imagine that Mars is under thorough

exploration, sometime in the first quarter of the twenty-first century.

However, there is no "Martian colony." There are no three-stage

rockets, no pressure-domes, no tractor-trailers, no human settlers.

Instead, there are hundreds of insect-sized robots, every one of

them equipped not merely with "true television," but something much

more advanced. They are equipped for *telepresence.* A human

operator can see what they see, hear what they hear, even guide them

about at will (granted, of course, that there is a steep transmission

lag). These micro-rovers, crammed with cheap microchips and laser

photo-optics, are so exquisitely monitored that one can actually *feel*

the Martian grit beneath their little scuttling claws. Piloting one of

these babies down the Valles Marineris, or perhaps some unknown

cranny of the Moon -- now *that* really feels like "exploration." If

they were cheap enough, you could dune-buggy them.

No one lives in space stations, in this scenario. Instead, our

entire solar system is saturated with cheap monitoring devices. There

are no "rockets" any more. Most of these robot surrogates weigh less

than a kilogram. They are fired into orbit by small rail-guns mounted

on high-flying aircraft. Or perhaps they're launched by laser-ignition:

ground-based heat-beams that focus on small reaction-chambers and

provide their thrust. They might even be literally shot into orbit by

Jules Vernian "space guns" that use the intriguing, dirt-cheap

technology of Gerald Bull's Iraqi "super-cannon." This wacky but

promising technique would be utterly impractical for launching human

beings, since the acceleration g-load would shatter every bone in their

bodies; but these little machines are *tough.*

And small robots have many other advantages. Unlike manned

craft, robots can go into harm's way: into Jupiter's radiation belts, or

into the shrapnel-heavy rings of Saturn, or onto the acid-bitten

smoldering surface of Venus. They stay on their missions,

operational, not for mere days or weeks, but for decades. They are

extensions, not of human population, but of human senses.

And because they are small and numerous, they should be

cheap. The entire point of this scenario is to create a new kind of

space-probe that is cheap, small, disposable, and numerous: as cheap

and disposable as their parent technologies, microchips and video,

while taking advantage of new materials like carbon-fiber, fiber-

optics, ceramic, and artificial diamond.

The core idea of this particular vision is "fast, cheap, and out of

control." Instead of gigantic, costly, ultra-high-tech, one-shot efforts

like NASA's Hubble Telescope (crippled by bad optics) or NASA's

Galileo (currently crippled by a flaw in its communications antenna)

these micro-rovers are cheap, and legion, and everywhere. They get

crippled every day; but it doesn't matter much; there are hundreds

more, and no one's life is at stake. People, even quite ordinary people,

*rent time on them* in much the same way that you would pay for

satellite cable-TV service. If you want to know what Neptune looks

like today, you just call up a data center and *have a look for

yourself.*

This is a concept that would truly involve "the public" in space

exploration, rather than the necessarily tiny elite of astronauts. This

is a potential benefit that we might derive from abandoning the

expensive practice of launching actual human bodies into space. We

might find a useful analogy in the computer revolution: "mainframe"

space exploration, run by a NASA elite in labcoats, is replaced by a

"personal" space exploration run by grad students and even hobbyists.

In this scenario, "space exploration" becomes similar to other

digitized, computer-assisted media environments: scientific

visualization, computer graphics, virtual reality, telepresence. The

solar system is saturated, not by people, but by *media coverage.

Outer space becomes *outer cyberspace.*

Whether this scenario is "realistic" isn't clear as yet. It's just a

science-fictional dream, a vision for the exploration of space:

*circumsolar telepresence.* As always, much depends on

circumstance, lucky accidents, and imponderables like political will.

What does seem clear, however, is that NASA's own current plans are

terribly far-fetched: they have outlived all contact with the political,

economic, social and even technical realities of the 1990s. There is no

longer any real point in shipping human beings into space in order to

wave flags.

"Exploring space" is not an "unrealistic" idea. That much, at

least, has already been proven. The struggle now is over why and

how and to what end. True, "exploring space" is not as "important"

as was the life-and-death Space Race struggle for Cold War pre-

eminence. Space science cannot realistically expect to command the

huge sums that NASA commanded in the service of American political

prestige. That era is simply gone; it's history now.

However: astronomy does count. There is a very deep and

genuine interest in these topics. An interest in the stars and planets is

not a fluke, it's not freakish. Astronomy is the most ancient of human

sciences. It's deeply rooted in the human psyche, has great historical

continuity, and is spread all over the world. It has its own

constituency, and if its plans were modest and workable, and played

to visible strengths, they might well succeed brilliantly.

The world doesn't actually need NASA's billions to learn about

our solar system. Real, honest-to-goodness "space exploration"

never got more than a fraction of NASA's budget in the first place.

Projects of this sort would no longer be created by gigantic

federal military-industrial bureaucracies. Micro-rover projects could

be carried out by universities, astronomy departments, and small-

scale research consortia. It would play from the impressive strengths

of the thriving communications and computer tech of the nineties,

rather than the dying, centralized, militarized, politicized rocket-tech

of the sixties.

The task at hand is to create a change in the climate of opinion

about the true potentials of "space exploration." Space exploration,

like the rest of us, grew up in the Cold War; like the rest of us, it must

now find a new way to live. And, as history has proven, science fiction

has a very real and influential role in space exploration. History

shows that true space exploration is not about budgets. It's about

vision. At its heart it has always been about vision.

Let's create the vision.

BUCKYMANIA

Carbon, like every other element on this planet, came to us from

outer space. Carbon and its compounds are well-known in galactic

gas-clouds, and in the atmosphere and core of stars, which burn

helium to produce carbon. Carbon is the sixth element in the periodic

table, and forms about two-tenths of one percent of Earth's crust.

Earth's biosphere (most everything that grows, moves, breathes,

photosynthesizes, or reads F&SF) is constructed mostly of

waterlogged carbon, with a little nitrogen, phosphorus and such for

leavening.

There are over a million known and catalogued compounds of

carbon: the study of these compounds, and their profuse and intricate

behavior, forms the major field of science known as organic

chemistry.

Since prehistory, "pure" carbon has been known to humankind

in three basic flavors. First, there's smut (lampblack or "amorphous

carbon"). Then there's graphite: soft, grayish-black, shiny stuff --

(pencil "lead" and lubricant). And third is that surpassing anomaly,

"diamond," which comes in extremely hard translucent crystals.

Smut is carbon atoms that are poorly linked. Graphite is carbon

atoms neatly linked in flat sheets. Diamond is carbon linked in strong,

regular, three-dimensional lattices: tetrahedra, that form ultrasolid

little carbon pyramids.

Today, however, humanity rejoices in possession of a fourth

and historically unprecedented form of carbon. Researchers have

created an entire class of these simon-pure carbon molecules, now

collectively known as the "fullerenes." They were named in August

1985, in Houston, Texas, in honor of the American engineer, inventor,

and delphically visionary philosopher, R. Buckminster Fuller.

"Buckminsterfullerene," or C60, is the best-known fullerene.

It's very round, the roundest molecule known to science. Sporting

what is technically known as "truncated icosahedral structure," C60 is

the most symmetric molecule possible in three-dimensional Euclidean

space. Each and every molecule of "Buckminsterfullerene" is a

hollow, geodesic sphere of sixty carbon atoms, all identically linked in

a spherical framework of twelve pentagons and twenty hexagons.

This molecule looks exactly like a common soccerball, and was

therefore nicknamed a "buckyball" by delighted chemists.

A free buckyball rotates merrily through space at one hundred

million revolutions per second. It's just over one nanometer across.

Buckminsterfullerene by the gross forms a solid crystal, is stable at

room temperature, and is an attractive mustard-yellow color. A heap

of crystallized buckyballs stack very much like pool balls, and are as

soft as graphite. It's thought that buckyballs will make good

lubricants -- something like molecular ball bearings.

When compressed, crystallized buckyballs squash and flatten

readily, down to about seventy percent of their volume. They then

refused to move any further and become extremely hard. Just *how*

hard is not yet established, but according to chemical theory,

compressed buckyballs may be considerably harder than diamond.

They may make good shock absorbers, or good armor.

But this is only the beginning of carbon's multifarious oddities in

the playful buckyball field. Because buckyballs are hollow, their

carbon framework can be wrapped around other, entirely different

atoms, forming neat molecular cages. This has already been

successfully done with certain metals, creating the intriguing new

class of "metallofullerites." Then there are buckyballs with a carbon or

two knocked out of the framework, and replaced with metal atoms.

This "doping" process yields a galaxy of so-called "dopeyballs." Some

of these dopeyballs show great promise as superconductors. Other

altered buckyballs seem to be organic ferromagnets.

A thin film of buckyballs can double the frequency of laser light

passing through it. Twisted or deformed buckyballs might act as

optical switches for future fiber-optic networks. Buckyballs with

dangling branches of nickel, palladium, or platinum may serve as new

industrial catalysts.

The electrical properties of buckyballs and their associated

compounds are very unusual, and therefore very promising. Pure C60

is an insulator. Add three potassium atoms, and it becomes a low-

temperature superconductor. Add three more potassium atoms, and it

becomes an insulator again! There's already excited talk in industry of

making electrical batteries out of buckyballs.

Then there are the "buckybabies:" C28, C32, C44, and C52. The

lumpy, angular buckybabies have received very little study to date,

and heaven only knows what they're capable of, especially when

doped, bleached, twisted, frozen or magnetized. And then there are

the *big* buckyballs: C240, C540, C960. Molecular models of these

monster buckyballs look like giant chickenwire beachballs.

There doesn't seem to be any limit to the upper size of a

buckyball. If wrapped around one another for internal support,

buckyballs can (at least theoretically) accrete like pearls. A truly

titanic buckyball might be big enough to see with the naked eye.

Conceivably, it might even be big enough to kick around on a playing

field, if you didn't mind kicking an anomalous entity with unknown

physical properties.

Carbon-fiber is a high-tech construction material which has

been seeing a lot of use lately in tennis rackets, bicycles, and high-

performance aircraft. It's already the strongest fiber known. This

makes the discovery of "buckytubes" even more striking. A buckytube

is carbon-fiber with a difference: it's a buckyball extruded into a long

continuous cylinder comprised of one single superstrong molecule.

C70, a buckyball cousin shaped like a rugby ball, seems to be

useful in producing high-tech films of artificial diamond. Then there

are "fuzzyballs" with sixty strands of hydrogen hair, "bunnyballs"

with twin ears of butylpyridine, flourinated "teflonballs" that may be

the slipperiest molecules ever produced.

This sudden wealth of new high-tech slang indicates the

potential riches of this new and multidisciplinary field of study, where

physics, electronics, chemistry and materials-science are all

overlapping, right now, in an exhilirating microsoccerball

scrimmage.

Today there are more than fifty different teams of scientists

investigating buckyballs and their relations, including industrial

heavy-hitters from AT&T, IBM and Exxon. SCIENCE magazine

voted buckminsterfullerene "Molecule of the Year" in 1991. Buckyball

papers have also appeared in NATURE, NEW SCIENTIST,

SCIENTIFIC AMERICAN, even FORTUNE and BUSINESS WEEK.

Buckyball breakthroughs are coming well-nigh every week, while the

fax machines sizzle in labs around the world. Buckyballs are strange,

elegant, beautiful, very intellectually sexy, and will soon be

commercially hot.

In chemical terms, the discovery of buckminsterfullerene -- a

carbon sphere -- may well rank with the discovery of the benzene ring

-- a carbon ring -- in the 19th century. The benzene ring (C6H6)

brought the huge field of aromatic chemistry into being, and with it a

enormous number of industrial applications.

But what was this "discovery," and how did it come about?

In a sense, like carbon itself, buckyballs also came to us from

outer space. Donald Huffman and Wolfgang Kratschmer were

astrophysicists studying interstellar soot. Huffman worked for the

University of Arizona in Tucson, Kratschmer for the Max Planck

Institute in Heidelberg. In 1982, these two gentlemen were

superheating graphite rods in a low-pressure helium atmosphere,

trying to replicate possible soot-making conditions in the atmosphere

of red-giant stars. Their experiment was run in a modest bell-jar

zapping apparatus about the size and shape of a washing-machine.

Among a great deal of black gunk, they actually manufactured

miniscule traces of buckminsterfullerene, which behaved oddly in their

spectrometer. At the time, however, they didn't realize what they

had.

In 1985, buckministerfullerene surfaced again, this time in a

high-tech laser-vaporization cluster-beam apparatus. Robert Curl

and Richard Smalley, two professors of chemistry at Rice University

in Houston, knew that a round carbon molecule was theoretically

possible. They even knew that it was likely to be yellow in color. And

in August 1985, they made a few nanograms of it, detected it with

mass spectrometers, and had the honor of naming it, along with their

colleagues Harry Kroto, Jim Heath and Sean O'Brien.

In 1985, however, there wasn't enough buckminsterfullerene

around to do much more than theorize about. It was "discovered,"

and named, and argued about in scientific journals, and was an

intriguing intellectual curiosity. But this exotic substance remained

little more than a lab freak.

And there the situation languished. But in 1988, Huffman and

Kratschmer, the astrophysicists, suddenly caught on: this "C60" from

the chemists in Houston, was probably the very same stuff they'd

made by a different process, back in 1982. Harry Kroto, who had

moved to the University of Sussex in the meantime, replicated their

results in his own machine in England, and was soon producing

enough buckminsterfullerene to actually weigh on a scale, and

measure, and purify!

The Huffman/Kratschmer process made buckminsterfullerene

by whole milligrams. Wow! Now the entire arsenal of modern

chemistry could be brought to bear: X-ray diffraction,

crystallography, nuclear magnetic resonance, chromatography. And

results came swiftly, and were published. Not only were buckyballs

real, they were weird and wonderful.

In 1990, the Rice team discovered a yet simpler method to make

buckyballs, the so-called "fullerene factory." In a thin helium

atmosphere inside a metal tank, a graphite rod is placed near a

graphite disk. Enough simple, brute electrical power is blasted

through the graphite to generate an electrical arc between the disk

and the tip of the rod. When the end of the rod boils off, you just crank

the stub a little closer and turn up the juice. The resultant exotic soot,

which collects on the metal walls of the chamber, is up to 45 percent

buckyballs.

In 1990, the buckyball field flung open its stadium doors for

anybody with a few gas-valves and enough credit for a big electric

bill. These buckyball "factories" sprang up all over the world in 1990

and '91. The "discovery" of buckminsterfullerene was not the big kick-

off in this particular endeavour. What really counted was the budget,

the simplicity of manufacturing. It wasn't the intellectual

breakthrough that made buckyballs a sport -- it was the cheap ticket in

through the gates. With cheap and easy buckyballs available, the

research scene exploded.

Sometimes Science, like other overglamorized forms of human

endeavor, marches on its stomach.

As I write this, pure buckyballs are sold commercially for about

$2000 a gram, but the market price is in free-fall. Chemists suggest

that buckmisterfullerene will be as cheap as aluminum some day soon

-- a few bucks a pound. Buckyballs will be a bulk commodity, like

oatmeal. You may even *eat* them some day -- they're not

poisonous, and they seem to offer a handy way to package certain

drugs.

Buckminsterfullerene may have been "born" in an interstellar

star-lab, but it'll become a part of everyday life, your life and my life,

like nylon, or latex, or polyester. It may become more famous, and

will almost certainly have far more social impact, than Buckminster

Fuller's own geodesic domes, those glamorously high-tech structures

of the 60s that were the prophetic vision for their molecule-size

counterparts.

This whole exciting buckyball scrimmage will almost certainly

bring us amazing products yet undreamt-of, everything from grease

to superhard steels. And, inevitably, it will bring a concomitant set of

new problems -- buckyball junk, perhaps, or bizarre new forms of

pollution, or sinister military applications. This is the way of the

world.

But maybe the most remarkable thing about this peculiar and

elaborate process of scientific development is that buckyballs never

were really "exotic" in the first place. Now that sustained attention

has been brought to bear on the phenomenon, it appears that

buckyballs are naturally present -- in tiny amounts, that is -- in almost

any sooty, smoky flame. Buckyballs fly when you light a candle, they

flew when Bogie lit a cigarette in "Casablanca," they flew when

Neanderthals roasted mammoth fat over the cave fire. Soot we knew

about, diamonds we prized -- but all this time, carbon, good ol'

Element Six, has had a shocking clandestine existence. The "secret"

was always there, right in the air, all around all of us.

But when you come right down to it, it doesn't really matter

how we found out about buckyballs. Accidents are not only fun, but

crucial to the so-called march of science, a march that often moves

fastest when it's stumbling down some strange gully that no one knew

existed. Scientists are human beings, and human beings are flexible:

not a hard, rigidly locked crystal like diamond, but a resilient network.

It's a legitimate and vital part of science to recognize the truth -- not

merely when looking for it with brows furrowed and teeth clenched,

but when tripping over it headlong.

Thanks to science, we did find out the truth. And now it's all

different. Because now we know!

THINK OF THE PRESTIGE

The science of rocketry, and the science of weaponry, are sister

sciences. It's been cynically said of German rocket scientist Wernher

von Braun that "he aimed at the stars, and hit London."

After 1945, Wernher von Braun made a successful transition to

American patronage and, eventually, to civilian space exploration.

But another ambitious space pioneer -- an American citizen -- was

not so lucky as von Braun, though his equal in scientific talent. His

story, by comparison, is little known.

Gerald Vincent Bull was born in March 9, 1928, in Ontario,

Canada. He died in 1990. Dr. Bull was the most brilliant artillery

scientist of the twentieth century. Bull was a prodigiously gifted

student, and earned a Ph.D. in aeronautical engineering at the age of

24.

Bull spent the 1950s researching supersonic aerodynamics in

Canada, personally handcrafting some of the most advanced wind-

tunnels in the world.

Bull's work, like that of his predecessor von Braun, had military

applications. Bull found patronage with the Canadian Armament

Research and Development Establishment (CARDE) and the

Canadian Defence Research Board.

However, Canada's military-industrial complex lacked the

panache, and the funding, of that of the United States. Bull, a

visionary and energetic man, grew impatient with what he considered

the pedestrian pace and limited imagination of the Canadians. As an

aerodynamics scientist for CARDE, Bull's salary in 1959 was only

$17,000. In comparison, in 1961 Bull earned $100,000 by consulting for

the Pentagon on nose-cone research. It was small wonder that by the

early 1960s, Bull had established lively professional relationships with

the US Army's Ballistics Research Laboratory (as well as the Army's

Redstone Arsenal, Wernher von Braun's own postwar stomping

grounds).

It was the great dream of Bull's life to fire cannon projectiles

from the earth's surface directly into outer space. Amazingly, Dr.

Bull enjoyed considerable success in this endeavor. In 1961, Bull

established Project HARP (High Altitude Research Project). HARP

was an academic, nonmilitary research program, funded by McGill

University in Montreal, where Bull had become a professor in the

mechanical engineering department. The US Army's Ballistic

Research Lab was a quiet but very useful co-sponsor of HARP; the US

Army was especially generous in supplying Bull with obsolete military

equipment, including cannon barrels and radar.

Project HARP found a home on the island of Barbados,

downrange of its much better-known (and vastly better-financed)

rival, Cape Canaveral. In Barbados, Bull's gigantic space-cannon

fired its projectiles out to an ocean splashdown, with little risk of

public harm. Its terrific boom was audible all over Barbados, but the

locals were much pleased at their glamorous link to the dawning

Space Age.

Bull designed a series of new supersonic shells known as the

"Martlets." The Mark II Martlets were cylindrical finned projectiles,

about eight inches wide and five feet six inches long. They weighed

475 pounds. Inside the barrel of the space-cannon, a Martlet was

surrounded by a precisely machined wooden casing known as a

"sabot." The sabot soaked up combustive energy as the projectile

flew up the space-cannon's sixteen-inch, 118-ft long barrel. As it

cleared the barrel, the sabot split and the precisely streamlined

Martlet was off at over a mile per second. Each shot produced a huge

explosion and a plume of fire gushing hundreds of feet into the sky.

The Martlets were scientific research craft. They were

designed to carry payloads of metallic chaff, chemical smoke, or

meteorological balloons. They sported telemetry antennas for tracing

the flight.

By the end of 1965, the HARP project had fired over a hundred

such missiles over fifty miles high, into the ionosphere -- the airless

fringes of space. In November 19, 1966, the US Army's Ballistics

Research Lab, using a HARP gun designed by Bull, fired a 185-lb

Martlet missile one hundred and eleven miles high. This was, and

remains, a world altitude record for any fired projectile. Bull now

entertained ambitious plans for a Martlet Mark IV, a rocket-assisted

projectile that would ignite in flight and drive itself into actual orbit.

Ballistically speaking, space cannon offer distinct advantages

over rockets. Rockets must lift, not only their own weight, but the

weight of their fuel and oxidizer. Cannon "fuel," which is contained

within the gunbarrel, offers far more explosive bang for the buck than

rocket fuel. Cannon projectiles are very accurate, thanks to the fixed

geometry of the gun-barrel. And cannon are far simpler and cheaper

than rockets.

There are grave disadvantages, of course. First, the payload

must be slender enough to fit into a gun-barrel. The most severe

drawback is the huge acceleration force of a cannon blast, which in the

case of Bull's exotic arsenal could top 10,000 Gs. This rules out

manned flights from the mouth of space-cannon. Jules Verne

overlooked this unpoetic detail when he wrote his prescient tale of

space artillery, FROM THE EARTH TO THE MOON (1865). (Dr Bull

was fascinated by Verne, and often spoke of Verne's science fiction as

one of the foremost inspirations of his youth.)

Bull was determined to put a cannon-round into orbit. This

burning desire of his was something greater than any merely

pragmatic or rational motive. The collapse of the HARP project in

1967 left Bull in command of his own fortunes. He reassembled the

wreckage of his odd academic/military career, and started a

commercial operation, "Space Research Corporation." In the years

to follow, Bull would try hard to sell his space-cannon vision to a

number of sponsors, including NATO, the Pentagon, Canada, China,

Israel, and finally, Iraq.

In the meantime, the Vietnam War was raging. Bull's

researches on projectile aerodynamics had made him, and his

company Space Reseach Corporation, into a hot military-industrial

property. In pursuit of space research, Bull had invented techniques

that lent much greater range and accuracy to conventional artillery

rounds. With Bull's ammunition, for instance, US Naval destroyers

would be able to cruise miles off the shore of North Vietnam,

destroying the best Russian-made shore batteries without any fear of

artillery retaliation. Bull's Space Research Corporation was

manufacturing the necessary long-range shells in Canada, but his lack

of American citizenship was a hindrance in the Pentagon arms trade.

Such was Dr. Bull's perceived strategic importance that this

hindrance was neatly avoided; with the sponsorship of Senator Barry

Goldwater, Bull became an American citizen by act of Congress. This

procedure was a rare honor, previously reserved only for Winston

Churchill and the Marquis de Lafayette.

Despite this Senatorial fiat, however, the Navy arms deal

eventually fell through. But although the US Navy scorned Dr. Bull's

wares, others were not so short-sighted. Bull's extended-range

ammunition, and the murderously brilliant cannon that he designed to

fire it, found ready markets in Egypt, Israel, Holland, Italy, Britain,

Canada, Venezuela, Chile, Thailand, Iran, South Africa, Austria and

Somalia.

Dr. Bull created a strange private reserve on the Canadian-

American border; a private arms manufactury with its own US and

Canadian customs units. This arrangement was very useful, since the

arms-export laws of the two countries differed, and SRC's military

products could be shipped-out over either national border at will. In

this distant enclave on the rural northern border of Vermont, the

arms genius built his own artillery range, his own telemetry towers

and launch-control buildings, his own radar tracking station,

workshops, and machine shops. At its height, the Space Research

Corporation employed over three hundred people at this site, and

boasted some $15 million worth of advanced equipment.

The downfall of HARP had left Bull disgusted with the

government-supported military-scientific establishment. He referred

to government researchers as "clowns" and "cocktail scientists," and

decided that his own future must lay in the vigorous world of free

enterprise. Instead of exploring the upper atmosphere, Bull

dedicated his ready intelligence to the refining of lethal munitions.

Bull would not sell to the Soviets or their client states, whom he

loathed; but he would sell to most anyone else. Bull's cannon are

credited with being of great help to Jonas Savimbi's UNITA war in

Angola; they were also extensively used by both sides in the Iran-Iraq

war.

Dr. Gerald V. Bull, Space Researcher, had become a

professional arms dealer. Dr. Bull was not a stellar success as an

arms dealer, because by all accounts he had no real head for business.

Like many engineers, Bull was obsessed not by entrepreneurial drive,

but by the exhilirating lure of technical achievement. The

atmosphere at Space Research Corporation was, by all accounts, very

collegial; Bull as professor, employees as cherished grad-students.

Bull's employees were fiercely loyal to him and felt that he was

brilliantly gifted and could accomplish anything.

SRC was never as great a commercial success as Bull's

technical genius merited. Bull stumbled badly in 1980. The Carter

Administration, annoyed by Bull's extensive deals with the South

African military, put Bull in prison for customs violation. This

punishment, rather than bringing Bull "to his senses," affected him

traumatically. He felt strongly that he had been singled out as a

political scapegoat to satisfy the hypocritical, left-leaning, anti-

apartheid bureaucrats in Washington. Bull spent seven months in an

American prison, reading extensively, and, incidentally, successfully

re-designing the prison's heating-plant. Nevertheless, the prison

experience left Bull embittered and cynical. While still in prison, Bull

was already accepting commercial approaches from the Communist

Chinese, who proved to be among his most avid customers.

After his American prison sentence ended, Bull abandoned his

strange enclave in the US-Canadian border to work full-time in

Brussels, Belgium. Space Research Corporation was welcomed there,

in Europe's foremost nexus of the global arms trade, a city where

almost anything goes in the way of merchandising war.

In November 1987, Bull was politely contacted in Brussels by the

Iraqi Embassy, and offered an all-expenses paid trip to Bagdad.

From 1980 to 1989, during their prolonged, lethal, and highly

inconclusive war with Iran, Saddam Hussein's regime had spent some

eighty billion dollars on weapons and weapons systems. Saddam

Hussein was especially fond of his Soviet-supplied "Scud" missiles,

which had shaken Iranian morale severely when fired into civilian

centers during the so-called "War of the Cities." To Saddam's mind,

the major trouble with his Scuds was their limited range and accuracy,

and he had invested great effort in gathering the tools and manpower

to improve the Iraqi art of rocketry.

The Iraqis had already bought many of Bull's 155-millimeter

cannon from the South Africans and the Austrians, and they were

most impressed. Thanks to Bull's design genius, the Iraqis actually

owned better, more accurate, and longer-range artillery than the

United States Army did.

Bull did not want to go to jail again, and was reluctant to break

the official embargo on arms shipments to Iraq. He told his would-be

sponsors so, in Bagdad, and the Iraqis were considerate of their

guest's qualms. To Bull's great joy, they took his idea of a peaceful

space cannon very seriously. "Think of the prestige," Bull suggested to

the Iraqi Minister of Industry, and the thought clearly intrigued the

Iraqi official.

The Israelis, in September 1988, had successfully launched their

own Shavit rocket into orbit, an event that had much impressed, and

depressed, the Arab League. Bull promised the Iraqis a launch system

that could place dozens, perhaps hundreds, of Arab satellites into

orbit. *Small* satellites, granted, and unmanned ones; but their

launches would cost as little as five thousand dollars each. Iraq

would become a genuine space power; a minor one by superpower

standards, but the only Arab space power.

And even small satellites were not just for show. Even a minor

space satellite could successfully perform certain surveillance

activities. The American military had proved the usefulness of spy

satellites to Saddam Hussein by passing him spysat intelligence during

worst heat of the Iran-Iraq war.

The Iraqis felt they would gain a great deal of widely

applicable, widely useful scientific knowledge from their association

with Bull, whether his work was "peaceful" or not. After all, it was

through peaceful research on Project HARP that Bull himself had

learned techniques that he had later sold for profit on the arms

market. The design of a civilian nose-cone, aiming for the stars, is

very little different from that of one descending with a supersonic

screech upon sleeping civilians in London.

For the first time in his life, Bull found himself the respected

client of a generous patron with vast resources -- and with an

imagination of a grandeur to match his own. By 1989, the Iraqis were

paying Bull and his company five million dollars a year to redesign

their field artillery, with much greater sums in the wings for "Project

Babylon" -- the Iraqi space-cannon. Bull had the run of ominous

weapons bunkers like the "Saad 16" missile-testing complex in north

Iraq, built under contract by Germans, and stuffed with gray-market

high-tech equipment from Tektronix, Scientific Atlanta and Hewlett-

Packard.

Project Babylon was Bull's grandest vision, now almost within

his grasp. The Iraqi space-launcher was to have a barrel five hundred

feet long, and would weigh 2,100 tons. It would be supported by a

gigantic concrete tower with four recoil mechanisms, these shock-

absorbers weighing sixty tons each. The vast, segmented cannon

would fire rocket-assisted projectiles the size of a phone booth, into

orbit around the Earth.

In August 1989, a smaller prototype, the so-called "Baby

Babylon," was constructed at a secret site in Jabal Hamrayn, in central

Iraq. "Baby Babylon" could not have put payloads into orbit, but it

would have had an international, perhaps intercontinental range.

The prototype blew up on its first test-firing.

The Iraqis continued undaunted on another prototype super-

gun, but their smuggling attempts were clumsy. Bull himself had little

luck in maintaining the proper discretion for a professional arms

dealer, as his own jailing had proved. When flattered, Bull talked;

and when he talked, he boasted.

Word began to leak out within the so-called "intelligence

community" that Bull was involved in something big; something to do

with Iraq and with missiles. Word also reached the Israelis, who were

very aware of Bull's scientific gifts, having dealt with him themselves,

extensively.

The Iraqi space cannon would have been nearly useless as a

conventional weapon. Five hundred feet long and completely

immobile, it would have been easy prey for any Israeli F-15. It would

have been impossible to hide, for any launch would thrown a column

of flame hundreds of feet into the air, a blazing signal for any spy

satellite or surveillance aircraft. The Babylon space cannon, faced

with determined enemies, could have been destroyed after a single

launch.

However, that single launch might well have served to dump a

load of nerve gas, or a nuclear bomb, onto any capital in the world.

Bull wanted Project Babylon to be entirely peaceful; despite his

rationalizations, he was never entirely at ease with military projects.

What Bull truly wanted from his Project Babylon was *prestige.* He

wanted the entire world to know that he, Jerry Bull, had created a

working space program, more or less all by himself. He had never

forgotten what it meant to world opinion to hear the Sputnik beeping

overhead.

For Saddam Hussein, Project Babylon was more than any

merely military weapon: it was a *political* weapon. The prestige

Iraq might gain from the success of such a visionary leap was worth

any number of mere cannon-fodder batallions. It was Hussein's

ambition to lead the Arab world; Bull's cannon was to be a symbol of

Iraqi national potency, a symbol that the long war with the Shi'ite

mullahs had not destroyed Saddam's ambitions for transcendant

greatness.

The Israelis, however, had already proven their willingness to

thwart Saddam Hussein's ambitions by whatever means necessary.

In 1981, they had bombed his Osirak nuclear reactor into rubble. In

1980, a Mossad hit-team had cut the throat of Iraqi nuclear scientist

Yayha El Meshad, in a Paris hotel room.

On March 22, 1990, Dr. Bull was surprised at the door of his

Brussels apartment. He was shot five times, in the neck and in the

back of the head, with a silenced 7.65 millimeter automatic pistol.

His assassin has never been found.

FOR FURTHER READING:

ARMS AND THE MAN: Dr. Gerald Bull, Iraq, and the Supergun by

William Lowther (McClelland- Bantam, Inc., Toronto, 1991)

BULL'S EYE: The Assassination and Life of Supergun Inventor

Gerald Bull by James Adams (Times Books, New York, 1992)

ARTIFICIAL LIFE

The new scientific field of study called "Artificial Life" can be

defined as "the attempt to abstract the logical form of life from its

material manifestation."

So far, so good. But what is life?

The basic thesis of "Artificial Life" is that "life" is best

understood as a complex systematic process. "Life" consists of

relationships and rules and interactions. "Life" as a property is

potentially separate from actual living creatures.

Living creatures (as we know them today, that is) are basically

made of wet organic substances: blood and bone, sap and cellulose,

chitin and ichor. A living creature -- a kitten, for instance -- is a

physical object that is made of molecules and occupies space and has

mass.

A kitten is indisputably "alive" -- but not because it has the

"breath of life" or the "vital impulse" somehow lodged inside its body.

We may think and talk and act as if the kitten "lives" because it has a

mysterious "cat spirit" animating its physical cat flesh. If we were

superstitious, we might even imagine that a healthy young cat had

*nine* lives. People have talked and acted just this way for millennia.

But from the point-of-view of Artificial Life studies, this is a

very halting and primitive way of conceptualizing what's actually

going on with a living cat. A kitten's "life" is a *process, * with

properties like reproduction, genetic variation, heredity, behavior,

learning, the possession of a genetic program, the expression of that

program through a physical body. "Life" is a thing that *does,* not a

thing that *is* -- life extracts energy from the environment, grows,

repairs damage, reproduces.

And this network of processes called "Life" can be picked apart,

and studied, and mathematically modelled, and simulated with

computers, and experimented upon -- outside of any creature's living

body.

"Artificial Life" is a very young field of study. The use of this

term dates back only to 1987, when it was used to describe a

conference in Los Alamos New Mexico on "the synthesis and

simulation of living systems." Artificial Life as a discipline is

saturated by computer-modelling, computer-science, and cybernetics.

It's conceptually similar to the earlier field of study called "Artificial

Intelligence." Artificial Intelligence hoped to extract the basic logical

structure of intelligence, to make computers "think." Artificial Life, by

contrast, hopes to make computers only about as "smart" as an ant --

but as "alive" as a swarming anthill.

Artificial Life as a discipline uses the computer as its primary

scientific instrument. Like telescopes and microscopes before them,

computers are making previously invisible aspects of the world

apparent to the human eye. Computers today are shedding light on

the activity of complex systems, on new physical principles such as

"emergent behavior," "chaos," and "self-organization."

For millennia, "Life" has been one of the greatest of

metaphysical and scientific mysteries, but now a few novel and

tentative computerized probes have been stuck into the fog. The

results have already proved highly intriguing.

Can a computer or a robot be alive? Can an entity which only

exists as a digital simulation be "alive"? If it looks like a duck, quacks

like a duck, waddles like a duck, but it in fact takes the form of pixels

on a supercomputer screen -- is it a duck? And if it's not a duck, then

what on earth is it? What exactly does a thing have to do and be

before we say it's "alive"?

It's surprisingly difficult to decide when something is "alive."

There's never been a definition of "life," whether scientific,

metaphysical, or theological, that has ever really worked. Life is not

a clean either/or proposition. Life comes on a kind of scale,

apparently, a kind of continuum -- maybe even, potentially, *several

different kinds of continuum.*

One might take a pragmatic, laundry-list approach to defining

life. To be "living," a thing must grow. Move. Reproduce. React to

its environment. Take in energy, excrete waste. Nourish itself, die,

and decay. Have a genetic code, perhaps, or be the result of a process

of evolution. But there are grave problems with all of these concepts.

All these things can be done today by machines or programs. And the

concepts themselves are weak and subject to contradiction and

paradox.

Are viruses "alive"? Viruses can thrive and reproduce, but not

by themselves -- they have to use a victim cell in order to manufacture

copies of themselves. Some dormant viruses can crystallize into a

kind of organic slag that's dead for all practical purposes, and can stay

that way indefinitely -- until the virus gets another chance at

infection, and then the virus comes seething back.

How about a frozen human embryo? It can be just as dormant

as a dormant virus, and certainly can't survive without a host, but it

can become a living human being. Some people who were once

frozen embryos may be reading this magazine right now! Is a frozen

embryo "alive" -- or is it just the *potential* for life, a genetic life-

program halted in mid-execution?

Bacteria are simple, as living things go. Most people however

would agree that germs are "alive." But there are many other entities

in our world today that act in lifelike fashion and are easily as

complex as germs, and yet we don't call them "alive" -- except

"metaphorically" (whatever *that* means).

How about a national government, for instance? A

government can grow and adapt and evolve. It's certainly a very

powerful entity that consumes resources and affects its environment

and uses enormous amounts of information. When people say "Long

Live France," what do they mean by that? Is the Soviet Union now

"dead"?

Amoebas aren't "mortal" and don't age -- they just go right on

splitting in half indefinitely. Does that mean that all amoebas are

actually pieces of one super-amoeba that's three billion years old?

And where's the "life" in an ant-swarm? Most ants in a swarm

never reproduce; they're sterile workers -- tools, peripherals,

hardware. All the individual ants in a nest, even the queen, can die

off one by one, but as long as new ants and new queens take their

place, the swarm itself can go on "living" for years without a hitch or a

stutter.

Questioning "life" in this way may seem so much nit-picking

and verbal sophistry. After all, one may think, people can easily tell

the difference between something living and dead just by having a

good long look at it. And in point of fact, this seems to be the single

strongest suit of "Artificial Life." It is very hard to look at a good

Artificial Life program in action without perceiving it as, somehow,

"alive."

Only living creatures perform the behavior known as

"flocking." A gigantic wheeling flock of cranes or flamingos is one of

the most impressive sights that the living world has to offer.

But the "logical form" of flocking can be abstracted from its

"material manifestation" in a flocking group of actual living birds.

"Flocking" can be turned into rules implemented on a computer. The

rules look like this:

1. Stay with the flock -- try to move toward where it seems

thickest.

2. Try to move at the same speed as the other local birds.

3. Don't bump into things, especially the ground or other birds.

In 1987, Craig Reynolds, who works for a computer-graphics

company called Symbolics, implemented these rules for abstract

graphic entities called "bird-oids" or "boids." After a bit of fine-

tuning, the result was, and is, uncannily realistic. The darn things

*flock!*

They meander around in an unmistakeably lifelike, lively,

organic fashion. There's nothing "mechanical" or "programmed-

looking" about their actions. They bumble and swarm. The boids in

the middle shimmy along contentedly, and the ones on the fringes tag

along anxiously jockeying for position, and the whole squadron hangs

together, and wheels and swoops and maneuvers, with amazing

grace. (Actually they're neither "anxious" nor "contented," but when

you see the boids behaving in this lifelike fashion, you can scarcely help

but project lifelike motives and intentions onto them.)

You might say that the boids simulate flocking perfectly -- but

according to the hard-dogma position of A-Life enthusiasts, it's not

"simulation" at all. This is real "flocking" pure and simple -- this is

exactly what birds actually do. Flocking is flocking -- it doesn't

matter if it's done by a whooping crane or a little computer-sprite.

Clearly the birdoids themselves aren't "alive" -- but it can be

argued, and is argued, that they're actually doing something that is a

genuine piece of the life process. In the words of scientist Christopher

Langton, perhaps the premier guru of A-Life: "The most important

thing to remember about A-Life is that the part that is artificial is not

the life, but the materials. Real things happen. We observe real

phenomena. It is real life in an artificial medium."

The great thing about studying flocking with boids, as opposed

to say whooping cranes, is that the Artificial Life version can be

experimented upon, in controlled and repeatable conditions. Instead

of just *observing* flocking, a life-scientist can now *do* flocking.

And not just flocks -- with a change in the parameters, you can study

"schooling" and "herding" as well.

The great hope of Artificial Life studies is that Artificial Life will

reveal previously unknown principles that directly govern life itself --

the principles that give life its mysterious complexity and power, its

seeming ability to defy probability and entropy. Some of these

principles, while still tentative, are hotly discussed in the field.

For instance: the principle of *bottom-up* initiative rather

than *top-down* orders. Flocking demonstrates this principle well.

Flamingos do not have blueprints. There is no squadron-leader

flamingo barking orders to all the other flamingos. Each flamingo

makes up its own mind. The extremely complex motion of a flock of

flamingos arises naturally from the interactions of hundreds of

independent birds. "Flocking" consists of many thousands of simple

actions and simple decisions, all repeated again and again, each

action and decision affecting the next in sequence, in an endless

systematic feedback.

This involves a second A-Life principle: *local* control rather

than *global* control. Each flamingo has only a vague notion of the

behavior of the flock as a whole. A flamingo simply isn't smart

enough to keep track of the entire "big picture," and in fact this isn't

even necessary. It's only necessary to avoid bumping the guys right

at your wingtips; you can safely ignore the rest.

Another principle: *simple* rules rather than *complex* ones.

The complexity of flocking, while real, takes place entirely outside of

the flamingo's brain. The individual flamingo has no mental

conception of the vast impressive aerial ballet in which it happens to

be taking part. The flamingo makes only simple decisions; it is never

required to make complex decisions requiring a lot of memory or

planning. *Simple* rules allow creatures as downright stupid as fish

to get on with the job at hand -- not only successfully, but swiftly and

gracefully.

And then there is the most important A-Life principle, also

perhaps the foggiest and most scientifically controversial:

*emergent* rather than *prespecified* behavior. Flamingos fly

from their roosts to their feeding grounds, day after day, year in year

out. But they will never fly there exactly the same way twice. They'll

get there all right, predictable as gravity; but the actual shape and

structure of the flock will be whipped up from scratch every time.

Their flying order is not memorized, they don't have numbered places

in line, or appointed posts, or maneuver orders. Their orderly

behavior simply *emerges,* different each time, in a ceaselessly

varying shuffle.

Ants don't have blueprints either. Ants have become the totem

animals of Artificial Life. Ants are so 'smart' that they have vastly

complex societies with actual *institutions* like slavery and and

agriculture and aphid husbandry. But an individual ant is a

profoundly stupid creature. Entomologists estimate that individual

ants have only fifteen to forty things that they can actually "do." But

if they do these things at the right time, to the right stimulus, and

change from doing one thing to another when the proper trigger

comes along, then ants as a group can work wonders.

There are anthills all over the world. They all work, but they're

all different; no two anthills are identical. That's because they're built

bottom-up and emergently. Anthills are built without any spark of

planning or intelligence. An ant may feel the vague instinctive need to

wall out the sunlight. It begins picking up bits of dirt and laying them

down at random. Other ants see the first ant at work and join in; this

is the A-Life principle known as "allelomimesis," imitating the others

(or rather not so much "imitating" them as falling mechanically into

the same instinctive pattern of behavior).

Sooner or later, a few bits of dirt happen to pile up together.

Now there's a wall. The ant wall-building sub-program kicks into

action. When the wall gets high enough, it's roofed over with dirt and

spit. Now there's a tunnel. Do it again and again and again, and the

structure can grow seven feet high, and be of such fantastic

complexity that to draw it on an architect's table would take years.

This emergent structure, "order out of chaos," "something out of

nothing" -- appears to be one of the basic "secrets of life."

These principles crop up again and again in the practice of life-

simulation. Predator-prey interactions. The effects of parasites and

viruses. Dynamics of population and evolution. These principles even

seem to apply to internal living processes, like plant growth and the

way a bug learns to walk. The list of applications for these principles

has gone on and on.

It's not hard to understand that many simple creatures, doing

simple actions that affect one another, can easily create a really big

mess. The thing that's *hard* to understand is that those same,

bottom-up, unplanned, "chaotic" actions can and do create living,

working, functional order and system and pattern. The process really

must be seen to be believed. And computers are the instruments that

have made us see it.

Most any computer will do. Oxford zoologist Richard

Dawkins has created a simple, popular Artificial Life program for

personal computers. It's called "The Blind Watchmaker," and

demonstrates the inherent power of Darwinian evolution to create

elaborate pattern and structure. The program accompanies Dr.

Dawkins' 1986 book of the same title (quite an interesting book, by the

way), but it's also available independently.

The Blind Watchmaker program creates patterns from little

black-and-white branching sticks, which develop according to very

simple rules. The first time you see them, the little branching sticks

seem anything but impressive. They look like this:

Fig 1. Ancestral A-Life Stick-Creature

After a pleasant hour with Blind Watchmaker, I myself produced

these very complex forms -- what Dawkins calls "Biomorphs."

Fig. 2 -- Six Dawkins Biomorphs

It's very difficult to look at such biomorphs without interpreting

them as critters -- *something* alive-ish, anyway. It seems that the

human eye is *trained by nature* to interpret the output of such a

process as "life-like." That doesn't mean it *is* life, but there's

definitely something *going on there.*

*What* is going on is the subject of much dispute. Is a

computer-simulation actually an abstracted part of life? Or is it

technological mimicry, or mechanical metaphor, or clever illusion?

We can model thermodynamic equations very well also, but an

equation isn't hot, it can't warm us or burn us. A perfect model of

heat isn't heat. We know how to model the flow of air on an

airplane's wings, but no matter how perfect our simulations are, they

don't actually make us fly. A model of motion isn't motion. Maybe

"Life" doesn't exist either, without that real-world carbon-and-water

incarnation. A-Life people have a term for these carbon-and-water

chauvinists. They call them "carbaquists."

Artificial Life maven Rodney Brooks designs insect-like robots

at MIT. Using A-Life bottom-up principles -- "fast, cheap, and out of

control" -- he is trying to make small multi-legged robots that can

behave as deftly as an ant. He and his busy crew of graduate students

are having quite a bit of success at it. And Brooks finds the struggle

over definitions beside the real point. He envisions a world in which

robots as dumb as insects are everywhere; dumb, yes, but agile and

successful and pragmatically useful. Brooks says: "If you want to

argue if it's living or not, fine. But if it's sitting there existing twenty-

four hours a day, three hundred sixty-five days of the year, doing

stuff which is tricky to do and doing it well, then I'm going to be

happy. And who cares what you call it, right?"

Ontological and epistemological arguments are never easily

settled. However, "Artificial Life," whether it fully deserves that term

or not, is at least easy to see, and rather easy to get your hands on.

"Blind Watchmaker" is the A-Life equivalent of using one's computer

as a home microscope and examining pondwater. Best of all, the

program costs only twelve bucks! It's cheap and easy to become an

amateur A-Life naturalist.

Because of the ubiquity of powerful computers, A-Life is

"garage-band science." The technology's out there for almost anyone

interested -- it's hacker-science. Much of A-Life practice basically

consists of picking up computers, pointing them at something

promising, and twiddling with the focus knobs until you see something

really gnarly. *Figuring out what you've seen* is the tough part, the

"real science"; this is where actual science, reproducible, falsifiable,

formal, and rigorous, parts company from the intoxicating glamor of

the intellectually sexy. But in the meantime, you have the contagious

joy and wonder of just *gazing at the unknown* the primal thrill of

discovery and exploration.

A lot has been written already on the subject of Artificial Life.

The best and most complete journalistic summary to date is Steven

Levy's brand-new book, ARTIFICIAL LIFE: THE QUEST FOR A NEW

CREATION (Pantheon Books 1992).

The easiest way for an interested outsider to keep up with this

fast-breaking field is to order books, videos, and software from an

invaluable catalog: "Computers In Science and Art," from Media

Magic. Here you can find the Proceedings of the first and second

Artificial Life Conferences, where the field's most influential papers,

discussions, speculations and manifestos have seen print.

But learned papers are only part of the A-Life experience. If

you can see Artificial Life actually demonstrated, you should seize the

opportunity. Computer simulation of such power and sophistication

is a truly remarkable historical advent. No previous generation had

the opportunity to see such a thing, much less ponder its significance.

Media Magic offers videos about cellular automata, virtual ants,

flocking, and other A-Life constructs, as well as personal software

"pocket worlds" like CA Lab, Sim Ant, and Sim Earth. This very

striking catalog is available free from Media Magic, P.O Box 507,

Nicasio CA 94946.

"INTERNET" [aka "A Short History of the Internet"]

Some thirty years ago, the RAND Corporation, America's

foremost Cold War think-tank, faced a strange strategic problem. How

could the US authorities successfully communicate after a nuclear

war?

Postnuclear America would need a command-and-control

network, linked from city to city, state to state, base to base. But no

matter how thoroughly that network was armored or protected, its

switches and wiring would always be vulnerable to the impact of

atomic bombs. A nuclear attack would reduce any

conceivable network to tatters.

And how would the network itself be commanded and

controlled? Any central authority, any network central citadel, would

be an obvious and immediate target for an enemy missile. The

center of the network would be the very first place to go.

RAND mulled over this grim puzzle in deep military secrecy,

and arrived at a daring solution. The RAND proposal (the brainchild

of RAND staffer Paul Baran) was made public in 1964. In the first

place, the network would *have no central authority.* Furthermore,

it would be *designed from the beginning to operate while

in tatters.*

The principles were simple. The network itself would be

assumed to be unreliable at all times. It would be designed from the

get-go to transcend its own unreliability. All the nodes in the network

would be equal in status to all other nodes, each node with its own

authority to originate, pass, and receive messages. The

messages themselves would be divided into packets, each packet

separately addressed. Each packet would begin at some specified

source node, and end at some other specified destination node. Each

packet would wind its way through the network on an individual

basis.

The particular route that the packet took would be unimportant.

Only final results would count. Basically, the packet would be tossed

like a hot potato from node to node to node, more or less in the

direction of its destination, until it ended up in the proper place. If

big pieces of the network had been blown away, that simply

wouldn't matter; the packets would still stay airborne, lateralled

wildly across the field by whatever nodes happened to survive. This

rather haphazard delivery system might be "inefficient" in the usual

sense (especially compared to, say, the telephone system) -- but it

would be extremely rugged.

During the 60s, this intriguing concept of a decentralized,

blastproof, packet-switching network was kicked around by RAND,

MIT and UCLA. The National Physical Laboratory in Great Britain set

up the first test network on these principles in 1968. Shortly

afterward, the Pentagon's Advanced Research Projects Agency decided

to fund a larger, more ambitious project in the USA. The nodes of the

network were to be high-speed supercomputers (or what passed for

supercomputers at the time). These were rare and valuable machines

which were in real need of good solid networking, for the sake of

national research-and-development projects.

In fall 1969, the first such node was installed in UCLA. By

December 1969, there were four nodes on the infant network, which

was named ARPANET, after its Pentagon sponsor.

The four computers could transfer data on dedicated high-

speed transmission lines. They could even be programmed remotely

from the other nodes. Thanks to ARPANET, scientists and researchers

could share one another's computer facilities by long-distance. This

was a very handy service, for computer-time was precious in the

early '70s. In 1971 there were fifteen nodes in ARPANET; by 1972,

thirty-seven nodes. And it was good.

By the second year of operation, however, an odd fact became

clear. ARPANET's users had warped the computer-sharing network

into a dedicated, high-speed, federally subsidized electronic post-

office. The main traffic on ARPANET was not long-distance computing.

Instead, it was news and personal messages. Researchers were using

ARPANET to collaborate on projects, to trade notes on work,

and eventually, to downright gossip and schmooze. People had their

own personal user accounts on the ARPANET computers, and their

own personal addresses for electronic mail. Not only were they using

ARPANET for person-to-person communication, but they were very

enthusiastic about this particular service -- far more enthusiastic than

they were about long-distance computation.

It wasn't long before the invention of the mailing-list, an

ARPANET broadcasting technique in which an identical message could

be sent automatically to large numbers of network subscribers.

Interestingly, one of the first really big mailing-lists was "SF-

LOVERS," for science fiction fans. Discussing science fiction on

the network was not work-related and was frowned upon by many

ARPANET computer administrators, but this didn't stop it from

happening.

Throughout the '70s, ARPA's network grew. Its decentralized

structure made expansion easy. Unlike standard corporate computer

networks, the ARPA network could accommodate many different

kinds of machine. As long as individual machines could speak the

packet-switching lingua franca of the new, anarchic network, their

brand-names, and their content, and even their ownership, were

irrelevant.

The ARPA's original standard for communication was known as

NCP, "Network Control Protocol," but as time passed and the technique

advanced, NCP was superceded by a higher-level, more sophisticated

standard known as TCP/IP. TCP, or "Transmission Control Protocol,"

converts messages into streams of packets at the source, then

reassembles them back into messages at the destination. IP, or

"Internet Protocol," handles the addressing, seeing to it that packets

are routed across multiple nodes and even across multiple networks

with multiple standards -- not only ARPA's pioneering NCP standard,

but others like Ethernet, FDDI, and X.25.

As early as 1977, TCP/IP was being used by other networks to

link to ARPANET. ARPANET itself remained fairly tightly controlled,

at least until 1983, when its military segment broke off and became

MILNET. But TCP/IP linked them all. And ARPANET itself, though it

was growing, became a smaller and smaller neighborhood amid the

vastly growing galaxy of other linked machines.

As the '70s and '80s advanced, many very different social

groups found themselves in possession of powerful computers. It was

fairly easy to link these computers to the growing network-of-

networks. As the use of TCP/IP became more common, entire other

networks fell into the digital embrace of the Internet, and

messily adhered. Since the software called TCP/IP was public-domain,

and the basic technology was decentralized and rather anarchic by its

very nature, it was difficult to stop people from barging in and

linking up somewhere-or-other. In point of fact, nobody *wanted* to

stop them from joining this branching complex of networks, which

came to be known as the "Internet."

Connecting to the Internet cost the taxpayer little or nothing,

since each node was independent, and had to handle its own financing

and its own technical requirements. The more, the merrier. Like the

phone network, the computer network became steadily more valuable

as it embraced larger and larger territories of people and resources.

A fax machine is only valuable if *everybody else* has a fax

machine. Until they do, a fax machine is just a curiosity. ARPANET,

too, was a curiosity for a while. Then computer-networking became

an utter necessity.

In 1984 the National Science Foundation got into the act,

through its Office of Advanced Scientific Computing. The new NSFNET

set a blistering pace for technical advancement, linking newer, faster,

shinier supercomputers, through thicker, faster links, upgraded and

expanded, again and again, in 1986, 1988, 1990. And other

government agencies leapt in: NASA, the National Institutes of Health,

the Department of Energy, each of them maintaining a digital satrapy

in the Internet confederation.

The nodes in this growing network-of-networks were divvied

up into basic varieties. Foreign computers, and a few American ones,

chose to be denoted by their geographical locations. The others were

grouped by the six basic Internet "domains": gov, mil, edu, com, org

and net. (Graceless abbreviations such as this are a standard

feature of the TCP/IP protocols.) Gov, Mil, and Edu denoted

governmental, military and educational institutions, which were, of

course, the pioneers, since ARPANET had begun as a high-tech

research exercise in national security. Com, however, stood

for "commercial" institutions, which were soon bursting into the

network like rodeo bulls, surrounded by a dust-cloud of eager

nonprofit "orgs." (The "net" computers served as gateways between

networks.)

ARPANET itself formally expired in 1989, a happy victim of its

own overwhelming success. Its users scarcely noticed, for ARPANET's

functions not only continued but steadily improved. The use of

TCP/IP standards for computer networking is now global. In 1971, a

mere twenty-one years ago, there were only four nodes in the

ARPANET network. Today there are tens of thousands of nodes in

the Internet, scattered over forty-two countries, with more coming

on-line every day. Three million, possibly four million people use

this gigantic mother-of-all-computer-networks.

The Internet is especially popular among scientists, and is

probably the most important scientific instrument of the late

twentieth century. The powerful, sophisticated access that it

provides to specialized data and personal communication

has sped up the pace of scientific research enormously.

The Internet's pace of growth in the early 1990s is spectacular,

almost ferocious. It is spreading faster than cellular phones, faster

than fax machines. Last year the Internet was growing at a rate of

twenty percent a *month.* The number of "host" machines with direct

connection to TCP/IP has been doubling every year since

1988. The Internet is moving out of its original base in military and

research institutions, into elementary and high schools, as well as into

public libraries and the commercial sector.

Why do people want to be "on the Internet?" One of the main

reasons is simple freedom. The Internet is a rare example of a true,

modern, functional anarchy. There is no "Internet Inc." There are

no official censors, no bosses, no board of directors, no stockholders.

In principle, any node can speak as a peer to any other node, as long

as it obeys the rules of the TCP/IP protocols, which are strictly

technical, not social or political. (There has been some struggle over

commercial use of the Internet, but that situation is changing as

businesses supply their own links).

The Internet is also a bargain. The Internet as a whole, unlike

the phone system, doesn't charge for long-distance service. And

unlike most commercial computer networks, it doesn't charge for

access time, either. In fact the "Internet" itself, which doesn't even

officially exist as an entity, never "charges" for anything. Each group

of people accessing the Internet is responsible for their own machine

and their own section of line.

The Internet's "anarchy" may seem strange or even unnatural,

but it makes a certain deep and basic sense. It's rather like the

"anarchy" of the English language. Nobody rents English, and nobody

owns English. As an English-speaking person, it's up to you to learn

how to speak English properly and make whatever use you please

of it (though the government provides certain subsidies to help you

learn to read and write a bit). Otherwise, everybody just sort of

pitches in, and somehow the thing evolves on its own, and somehow

turns out workable. And interesting. Fascinating, even. Though a lot

of people earn their living from using and exploiting and teaching

English, "English" as an institution is public property, a public good.

Much the same goes for the Internet. Would English be improved if

the "The English Language, Inc." had a board of directors and a chief

executive officer, or a President and a Congress? There'd probably be

a lot fewer new words in English, and a lot fewer new ideas.

People on the Internet feel much the same way about their own

institution. It's an institution that resists institutionalization. The

Internet belongs to everyone and no one.

Still, its various interest groups all have a claim. Business

people want the Internet put on a sounder financial footing.

Government people want the Internet more fully regulated.

Academics want it dedicated exclusively to scholarly research.

Military people want it spy-proof and secure. And so on and so on.

All these sources of conflict remain in a stumbling balance

today, and the Internet, so far, remains in a thrivingly anarchical

condition. Once upon a time, the NSFnet's high-speed, high-capacity

lines were known as the "Internet Backbone," and their owners could

rather lord it over the rest of the Internet; but today there are

"backbones" in Canada, Japan, and Europe, and even privately owned

commercial Internet backbones specially created for carrying business

traffic. Today, even privately owned desktop computers can become

Internet nodes. You can carry one under your arm. Soon, perhaps, on

your wrist.

But what does one *do* with the Internet? Four things,

basically: mail, discussion groups, long-distance computing, and file

transfers.

Internet mail is "e-mail," electronic mail, faster by several

orders of magnitude than the US Mail, which is scornfully known by

Internet regulars as "snailmail." Internet mail is somewhat like fax.

It's electronic text. But you don't have to pay for it (at least not

directly), and it's global in scope. E-mail can also send software and

certain forms of compressed digital imagery. New forms of mail are in

the works.

The discussion groups, or "newsgroups," are a world of their

own. This world of news, debate and argument is generally known as

"USENET. " USENET is, in point of fact, quite different from the

Internet. USENET is rather like an enormous billowing crowd of

gossipy, news-hungry people, wandering in and through the

Internet on their way to various private backyard barbecues.

USENET is not so much a physical network as a set of social

conventions. In any case, at the moment there are some 2,500

separate newsgroups on USENET, and their discussions generate about

7 million words of typed commentary every single day. Naturally

there is a vast amount of talk about computers on USENET, but the

variety of subjects discussed is enormous, and it's growing larger all

the time. USENET also distributes various free electronic journals and

publications.

Both netnews and e-mail are very widely available, even

outside the high-speed core of the Internet itself. News and e-mail

are easily available over common phone-lines, from Internet fringe-

realms like BITnet, UUCP and Fidonet. The last two Internet services,

long-distance computing and file transfer, require what is known as

"direct Internet access" -- using TCP/IP.

Long-distance computing was an original inspiration for

ARPANET and is still a very useful service, at least for some.

Programmers can maintain accounts on distant, powerful computers,

run programs there or write their own. Scientists can make use of

powerful supercomputers a continent away. Libraries offer their

electronic card catalogs for free search. Enormous CD-ROM catalogs

are increasingly available through this service. And there are

fantastic amounts of free software available.

File transfers allow Internet users to access remote machines

and retrieve programs or text. Many Internet computers -- some

two thousand of them, so far -- allow any person to access them

anonymously, and to simply copy their public files, free of charge.

This is no small deal, since entire books can be transferred through

direct Internet access in a matter of minutes. Today, in 1992, there

are over a million such public files available to anyone who asks for

them (and many more millions of files are available to people with

accounts). Internet file-transfers are becoming a new form of

publishing, in which the reader simply electronically copies the work

on demand, in any quantity he or she wants, for free. New Internet

programs, such as "archie," "gopher," and "WAIS," have been

developed to catalog and explore these enormous archives of

material.

The headless, anarchic, million-limbed Internet is spreading like

bread-mold. Any computer of sufficient power is a potential spore

for the Internet, and today such computers sell for less than $2,000

and are in the hands of people all over the world. ARPA's network,

designed to assure control of a ravaged society after a nuclear

holocaust, has been superceded by its mutant child the Internet,

which is thoroughly out of control, and spreading exponentially

through the post-Cold War electronic global village. The spread of

the Internet in the 90s resembles the spread of personal

computing in the 1970s, though it is even faster and perhaps more

important. More important, perhaps, because it may give those

personal computers a means of cheap, easy storage and access that is

truly planetary in scale.

The future of the Internet bids fair to be bigger and

exponentially faster. Commercialization of the Internet is a very hot

topic today, with every manner of wild new commercial information-

service promised. The federal government, pleased with an unsought

success, is also still very much in the act. NREN, the National Research

and Education Network, was approved by the US Congress in fall

1991, as a five-year, $2 billion project to upgrade the Internet

"backbone." NREN will be some fifty times faster than the fastest

network available today, allowing the electronic transfer of the entire

Encyclopedia Britannica in one hot second. Computer networks

worldwide will feature 3-D animated graphics, radio and cellular

phone-links to portable computers, as well as fax, voice, and high-

definition television. A multimedia global circus!

Or so it's hoped -- and planned. The real Internet of the

future may bear very little resemblance to today's plans. Planning

has never seemed to have much to do with the seething, fungal

development of the Internet. After all, today's Internet bears

little resemblance to those original grim plans for RAND's post-

holocaust command grid. It's a fine and happy irony.

How does one get access to the Internet? Well -- if you don't

have a computer and a modem, get one. Your computer can act as a

terminal, and you can use an ordinary telephone line to connect to an

Internet-linked machine. These slower and simpler adjuncts to the

Internet can provide you with the netnews discussion groups and

your own e-mail address. These are services worth having -- though

if you only have mail and news, you're not actually "on the Internet"

proper.

If you're on a campus, your university may have direct

"dedicated access" to high-speed Internet TCP/IP lines. Apply for an

Internet account on a dedicated campus machine, and you may be

able to get those hot-dog long-distance computing and file-transfer

functions. Some cities, such as Cleveland, supply "freenet"

community access. Businesses increasingly have Internet access, and

are willing to sell it to subscribers. The standard fee is about $40 a

month -- about the same as TV cable service.

As the Nineties proceed, finding a link to the Internet will

become much cheaper and easier. Its ease of use will also improve,

which is fine news, for the savage UNIX interface of TCP/IP leaves

plenty of room for advancements in user-friendliness. Learning the

Internet now, or at least learning about it, is wise. By the

turn of the century, "network literacy," like "computer literacy"

before it, will be forcing itself into the very texture of your life.

For Further Reading:

The Whole Internet Catalog & User's Guide by Ed Krol. (1992) O'Reilly

and Associates, Inc. A clear, non-jargonized introduction to the

intimidating business of network literacy. Many computer-

documentation manuals attempt to be funny. Mr. Krol's book is

*actually* funny.

The Matrix: Computer Networks and Conferencing Systems Worldwide.

by John Quarterman. Digital Press: Bedford, MA. (1990) Massive and

highly technical compendium detailing the mind-boggling scope and

complexity of our newly networked planet.

The Internet Companion by Tracy LaQuey with Jeanne C. Ryer (1992)

Addison Wesley. Evangelical etiquette guide to the Internet featuring

anecdotal tales of life-changing Internet experiences. Foreword by

Senator Al Gore.

Zen and the Art of the Internet: A Beginner's Guide by Brendan P.

Kehoe (1992) Prentice Hall. Brief but useful Internet guide with

plenty of good advice on useful machines to paw over for data. Mr

Kehoe's guide bears the singularly wonderful distinction of being

available in electronic form free of charge. I'm doing the same

with all my F&SF Science articles, including, of course, this one. My

own Internet address is bruces@well.sf.ca.us.

"Magnetic Vision"

Here on my desk I have something that can only be described as

miraculous. It's a big cardboard envelope with nine thick sheets of

black plastic inside, and on these sheets are pictures of my own brain.

These images are "MRI scans" -- magnetic resonance imagery from

a medical scanner.

These are magnetic windows into the lightless realm inside my

skull. The meat, bone, and various gristles within my head glow gently

in crisp black-and-white detail. There's little of the foggy ghostliness

one sees with, say, dental x-rays. Held up against a bright light, or

placed on a diagnostic light table, the dark plastic sheets reveal veins,

arteries, various odd fluid-stuffed ventricles, and the spongey wrinkles

of my cerebellum. In various shots, I can see the pulp within my own

teeth, the roots of my tongue, the boney caverns of my sinuses, and the

nicely spherical jellies that are my two eyeballs. I can see that the

human brain really does come in two lobes and in three sections, and

that it has gray matter and white matter. The brain is a big whopping

gland, basically, and it fills my skull just like the meat of a walnut.

It's an odd experience to look long and hard at one's own brain.

Though it's quite a privilege to witness this, it's also a form of

narcissism without much historical parallel. Frankly, I don't think I

ever really believed in my own brain until I saw these images. At least,

I never truly comprehended my brain as a tangible physical organ, like

a knuckle or a kneecap. And yet here is the evidence, laid out

irrefutably before me, pixel by monochrome pixel, in a large variety of

angles and in exquisite detail. And I'm told that my brain is quite

healthy and perfectly normal -- anatomically at least. (For a science

fiction writer this news is something of a letdown.)

The discovery of X-rays in 1895, by Wilhelm Roentgen, led to the

first technology that made human flesh transparent. Nowadays, X-rays

can pierce the body through many different angles to produce a

graphic three-dimensional image. This 3-D technique, "Computerized

Axial Tomography" or the CAT-scan, won a Nobel Prize in 1979 for its

originators, Godfrey Hounsfield and Allan Cormack.

Sonography uses ultrasound to study human tissue through its

reflection of high-frequency vibration: sonography is a sonic window.

Magnetic resonance imaging, however, is a more sophisticated

window yet. It is rivalled only by the lesser-known and still rather

experimental PET-scan, or Positron Emission Tomography. PET-

scanning requires an injection of radioactive isotopes into the body so

that their decay can be tracked within human tissues. Magnetic

resonance, though it is sometimes known as Nuclear Magnetic

Resonance, does not involve radioactivity.

The phenomenon of "nuclear magnetic resonance" was

discovered in 1946 by Edward Purcell of Harvard, and Felix Block of

Stanford. Purcell and Block were working separately, but published

their findings within a month of one another. In 1952, Purcell and

Block won a joint Nobel Prize for their discovery.

If an atom has an odd number of protons and neutrons, it will

have what is known as a "magnetic moment:" it will spin, and its axis

will tilt in a certain direction. When that tilted nucleus is put into a

magnetic field, the axis of the tilt will change, and the nucleus will also

wobble at a certain speed. If radio waves are then beamed at the

wobbling nucleus at just the proper wavelength, they will cause the

wobbling to intensify -- this is the "magnetic resonance" phenomenon.

The resonant frequency is known as the Larmor frequency, and the

Larmor frequencies vary for different atoms.

Hydrogen, for instance, has a Larmor frequency of 42.58

megahertz. Hydrogen, which is a major constituent of water and of

carbohydrates such as fat, is very common in the human body. If radio

waves at this Larmor frequency are beamed into magnetized hydrogen

atoms, the hydrogen nuclei will absorb the resonant energy until they

reach a state of excitation. When the beam goes off, the hydrogen

nuclei will relax again, each nucleus emitting a tiny burst of radio

energy as it returns to its original state. The nuclei will also relax at

slightly different rates, depending on the chemical circumstances

around the hydrogen atom. Hydrogen behaves differently in different

kinds of human tissue. Those relaxation bursts can be detected, and

timed, and mapped.

The enormously powerful magnetic field within an MRI machine

can permeate the human body; but the resonant Larmor frequency is

beamed through the body in thin, precise slices. The resulting images

are neat cross-sections through the body. Unlike X-rays, magnetic

resonance doesn't ionize and possibly damage human cells. Instead, it

gently coaxes information from many different types of tissue, causing

them to emit tell-tale signals about their chemical makeup. Blood, fat,

bones, tendons, all emit their own characteristics, which a computer

then reassembles as a graphic image on a computer screen, or prints

out on emulsion-coated plastic sheets.

An X-ray is a marvelous technology, and a CAT-scan more

marvelous yet. But an X-ray does have limits. Bones cast shadows in X-

radiation, making certain body areas opaque or difficult to read. And X-

ray images are rather stark and anatomical; an X-ray image cannot

even show if the patient is alive or dead. An MRI scan, on the other

hand, will reveal a great deal about the composition and the health of

living tissue. For instance, tumor cells handle their fluids differently

than normal tissue, giving rise to a slightly different set of signals. The

MRI machine itself was originally invented as a cancer detector.

After the 1946 discovery of magnetic resonance, MRI techniques

were used for thirty years to study small chemical samples. However, a

cancer researcher, Dr. Raymond Damadian, was the first to build an MRI

machine large enough and sophisticated enough to scan an entire

human body, and then produce images from that scan. Many scientists,

most of them even, believed and said that such a technology was decades

away, or even technically impossible. Damadian had a tough,

prolonged struggle to find funding for for his visionary technique, and

he was often dismissed as a zealot, a crackpot, or worse. Damadian's

struggle and eventual triumph is entertainingly detailed in his 1985

biography, A MACHINE CALLED INDOMITABLE.

Damadian was not much helped by his bitter and public rivalry

with his foremost competitor in the field, Paul Lauterbur. Lauterbur,

an industrial chemist, was the first to produce an actual magnetic-

resonance image, in 1973. But Damadian was the more technologically

ambitious of the two. His machine, "Indomitable," (now in the

Smithsonian Museum) produced the first scan of a human torso, in 1977.

(As it happens, it was Damadian's own torso.) Once this proof-of-

concept had been thrust before a doubting world, Damadian founded a

production company, and became the father of the MRI scanner

industry.

By the end of the 1980s, medical MRI scanning had become a

major enterprise, and Damadian had won the National Medal of

Technology, along with many other honors. As MRI machines spread

worldwide, the market for CAT-scanning began to slump in comparison.

Today, MRI is a two-billion dollar industry, and Dr Damadian and his

company, Fonar Corporation, have reaped the fruits of success. (Some

of those fruits are less sweet than others: today Damadian and Fonar

Corp. are suing Hitachi and General Electric in federal court, for

alleged infringement of Damadian's patents.)

MRIs are marvelous machines -- perhaps, according to critics, a

little too marvelous. The magnetic fields emitted by MRIs are extremely

strong, strong enough to tug wheelchairs across the hospital floor, to

wipe the data off the magnetic strips in credit cards, and to whip a

wrench or screwdriver out of one's grip and send it hurtling across the

room. If the patient has any metal imbedded in his skin -- welders and

machinists, in particular, often do have tiny painless particles of

shrapnel in them -- then these bits of metal will be wrenched out of the

patient's flesh, producing a sharp bee-sting sensation. And in the

invisible grip of giant magnets, heart pacemakers can simply stop.

MRI machines can weigh ten, twenty, even one hundred tons.

And they're big -- the scanning cavity, in which the patient is inserted,

is about the size and shape of a sewer pipe, but the huge plastic hull

surrounding that cavity is taller than a man and longer than a plush

limo. A machine of that enormous size and weight cannot be moved

through hospital doors; instead, it has to be delivered by crane, and its

shelter constructed around it. That shelter must not have any iron

construction rods in it or beneath its floor, for obvious reasons. And yet

that floor had better be very solid indeed.

Superconductive MRIs present their own unique hazards. The

superconductive coils are supercooled with liquid helium.

Unfortunately there's an odd phenomenon known as "quenching," in

which a superconductive magnet, for reasons rather poorly understood,

will suddenly become merely-conductive. When a "quench" occurs, an

enormous amount of electrical energy suddenly flashes into heat,

which makes the liquid helium boil violently. The MRI's technicians

might be smothered or frozen by boiling helium, so it has to be vented

out the roof, requiring the installation of specialized vent-stacks.

Helium leaks, too, so it must be resupplied frequently, at considerable

expense.

The MRI complex also requires expensive graphic-processing

computers, CRT screens, and photographic hard-copy devices. Some

scanners feature elaborate telecommunications equipment. Like the

giant scanners themselves, all these associated machines require

power-surge protectors, line conditioners, and backup power supplies.

Fluorescent lights, which produce radio-frequency noise pollution, are

forbidden around MRIs. MRIs are also very bothered by passing CB

radios, paging systems, and ambulance transmissions. It is generally

considered a good idea to sheathe the entire MRI cubicle (especially the

doors, windows, electrical wiring, and plumbing) in expensive, well-

grounded sheet-copper.

Despite all these drawbacks, the United States today rejoices in

possession of some two thousand MRI machines. (There are hundreds in

other countries as well.) The cheaper models cost a solid million dollars

each; the top-of-the-line models, two million. Five million MRI scans

were performed in the United States last year, at prices ranging from

six hundred dollars, to twice that price and more.

In other words, in 1991 alone, Americans sank some five billion

dollars in health care costs into the miraculous MRI technology.

Today America's hospitals and diagnostic clinics are in an MRI

arms race. Manufacturers constantly push new and improved machines

into the market, and other hospitals feel a dire need to stay with the

state-of-the-art. They have little choice in any case, for the balky,

temperamental MRI scanners wear out in six years or less, even when

treated with the best of care.

Patients have little reason to refuse an MRI test, since insurance

will generally cover the cost. MRIs are especially good for testing for

neurological conditions, and since a lot of complaints, even quite minor

ones, might conceivably be neurological, a great many MRI scans are

performed. The tests aren't painful, and they're not considered risky.

Having one's tissues briefly magnetized is considered far less risky than

the fairly gross ionization damage caused by X-rays. The most common

form of MRI discomfort is simple claustrophobia. MRIs are as narrow as

the grave, and also very loud, with sharp mechanical clacking and

buzzing.

But the results are marvels to behold, and MRIs have clearly

saved many lives. And the tests will eliminate some potential risks to

the patient, and put the physician on surer ground with his diagnosis.

So why not just go ahead and take the test?

MRIs have gone ahead boldly. Unfortunately, miracles rarely

come cheap. Today the United States spends thirteen percent of its Gross

National Product on health care, and health insurance costs are

drastically outstripping the rate of inflation.

High-tech, high-cost resources such as MRIs generally go to to

the well-to-do and the well-insured. This practice has sad

repercussions. While some lives are saved by technological miracles --

and this is a fine thing -- other lives are lost, that might have been

rescued by fairly cheap and common public-health measures, such as

better nutrition, better sanitation, or better prenatal care. As advanced

nations go, the United States a rather low general life expectancy, and a

quite bad infant-death rate; conspicuously worse, for instance, than

Italy, Japan, Germany, France, and Canada.

MRI may be a true example of a technology genuinely ahead of

its time. It may be that the genius, grit, and determination of Raymond

Damadian brought into the 1980s a machine that might have been better

suited to the technical milieu of the 2010s. What MRI really requires for

everyday workability is some cheap, simple, durable, powerful

superconductors. Those are simply not available today, though they

would seem to be just over the technological horizon. In the meantime,

we have built thousands of magnetic windows into the body that will do

more or less what CAT-scan x-rays can do already. And though they do

it better, more safely, and more gently than x-rays can, they also do it

at a vastly higher price.

Damadian himself envisioned MRIs as a cheap mass-produced

technology. "In ten to fifteen years," he is quoted as saying in 1985,

"we'll be able to step into a booth -- they'll be in shopping malls or

department stores -- put a quarter in it, and in a minute it'll say you

need some Vitamin A, you have some bone disease over here, your blood

pressure is a touch high, and keep a watch on that cholesterol." A

thorough medical checkup for twenty-five cents in 1995! If one needed

proof that Raymond Damadian was a true visionary, one could find it

here.

Damadian even envisioned a truly advanced MRI machine

capable of not only detecting cancer, but of killing cancerous cells

outright. These machines would excite not hydrogen atoms, but

phosphorus atoms, common in cancer-damaged DNA. Damadian

speculated that certain Larmor frequencies in phosphorus might be

specific to cancerous tissue; if that were the case, then it might be

possible to pump enough energy into those phosphorus nuclei so that

they actually shivered loose from the cancer cell's DNA, destroying the

cancer cell's ability to function, and eventually killing it.

That's an amazing thought -- a science-fictional vision right out

of the Gernback Continuum. Step inside the booth -- drop a quarter --

and have your incipient cancer not only diagnosed, but painlessly

obliterated by invisible Magnetic Healing Rays.

Who the heck could believe a visionary scenario like that?

Some things are unbelievable until you see them with your own

eyes. Until the vision is sitting right there in front of you. Where it

can no longer be denied that they're possible.

A vision like the inside of your own brain, for instance.

SUPERGLUE

This is the Golden Age of Glue.

For thousands of years, humanity got by with natural glues like

pitch, resin, wax, and blood; products of hoof and hide and treesap

and tar. But during the past century, and especially during the past

thirty years, there has been a silent revolution in adhesion.

This stealthy yet steady technological improvement has been

difficult to fully comprehend, for glue is a humble stuff, and the

better it works, the harder it is to notice. Nevertheless, much of the

basic character of our everyday environment is now due to advanced

adhesion chemistry.

Many popular artifacts from the pre-glue epoch look clunky

and almost Victorian today. These creations relied on bolts, nuts,

rivets, pins, staples, nails, screws, stitches, straps, bevels, knobs, and

bent flaps of tin. No more. The popular demand for consumer

objects ever lighter, smaller, cheaper, faster and sleeker has led to

great changes in the design of everyday things.

Glue determines much of the difference between our

grandparent's shoes, with their sturdy leather soles, elaborate

stitching, and cobbler's nails, and the eerie-looking modern jogging-

shoe with its laminated plastic soles, fabric uppers and sleek foam

inlays. Glue also makes much of the difference between the big

family radio cabinet of the 1940s and the sleek black hand-sized

clamshell of a modern Sony Walkman.

Glue holds this very magazine together. And if you happen to

be reading this article off a computer (as you well may), then you

are even more indebted to glue; modern microelectronic assembly

would be impossible without it.

Glue dominates the modern packaging industry. Glue also has

a strong presence in automobiles, aerospace, electronics, dentistry,

medicine, and household appliances of all kinds. Glue infiltrates

grocery bags, envelopes, books, magazines, labels, paper cups, and

cardboard boxes; there are five different kinds of glue in a common

filtered cigarette. Glue lurks invisibly in the structure of our

shelters, in ceramic tiling, carpets, counter tops, gutters, wall siding,

ceiling panels and floor linoleum. It's in furniture, cooking utensils,

and cosmetics. This galaxy of applications doesn't even count the

vast modern spooling mileage of adhesive tapes: package tape,

industrial tape, surgical tape, masking tape, electrical tape, duct tape,

plumbing tape, and much, much more.

Glue is a major industrial industry and has been growing at

twice the rate of GNP for many years, as adhesives leak and stick

into areas formerly dominated by other fasteners. Glues also create

new markets all their own, such as Post-it Notes (first premiered in

April 1980, and now omnipresent in over 350 varieties).

The global glue industry is estimated to produce about twelve

billion pounds of adhesives every year. Adhesion is a $13 billion

market in which every major national economy has a stake. The

adhesives industry has its own specialty magazines, such as

Adhesives Age andSAMPE Journal; its own trade groups, like the

Adhesives Manufacturers Association, The Adhesion Society, and the

Adhesives and Sealant Council; and its own seminars, workshops and

technical conferences. Adhesives corporations like 3M, National

Starch, Eastman Kodak, Sumitomo, and Henkel are among the world's

most potent technical industries.

Given all this, it's amazing how little is definitively known

about how glue actually works -- the actual science of adhesion.

There are quite good industrial rules-of-thumb for creating glues;

industrial technicians can now combine all kinds of arcane

ingredients to design glues with well-defined specifications:

qualities such as shear strength, green strength, tack, electrical

conductivity, transparency, and impact resistance. But when it

comes to actually describing why glue is sticky, it's a different

matter, and a far from simple one.

A good glue has low surface tension; it spreads rapidly and

thoroughly, so that it will wet the entire surface of the substrate.

Good wetting is a key to strong adhesive bonds; bad wetting leads

to problems like "starved joints," and crannies full of trapped air,

moisture, or other atmospheric contaminants, which can weaken the

bond.

But it is not enough just to wet a surface thoroughly; if that

were the case, then water would be a glue. Liquid glue changes

form; it cures, creating a solid interface between surfaces that

becomes a permanent bond.

The exact nature of that bond is pretty much anybody's guess.

There are no less than four major physico-chemical theories about

what makes things stick: mechanical theory, adsorption theory,

electrostatic theory and diffusion theory. Perhaps molecular strands

of glue become physically tangled and hooked around irregularities

in the surface, seeping into microscopic pores and cracks. Or, glue

molecules may be attracted by covalent bonds, or acid-base

interactions, or exotic van der Waals forces and London dispersion

forces, which have to do with arcane dipolar resonances between

magnetically imbalanced molecules. Diffusion theorists favor the

idea that glue actually blends into the top few hundred molecules of

the contact surface.

Different glues and different substrates have very different

chemical constituents. It's likely that all of these processes may have

something to do with the nature of what we call "stickiness" -- that

everybody's right, only in different ways and under different

circumstances.

In 1989 the National Science Foundation formally established

the Center for Polymeric Adhesives and Composites. This Center's

charter is to establish "a coherent philosophy and systematic

methodology for the creation of new and advanced polymeric

adhesives" -- in other words, to bring genuine detailed scientific

understanding to a process hitherto dominated by industrial rules of

thumb. The Center has been inventing new adhesion test methods

involving vacuum ovens, interferometers, and infrared microscopes,

and is establishing computer models of the adhesion process. The

Center's corporate sponsors -- Amoco, Boeing, DuPont, Exxon,

Hoechst Celanese, IBM, Monsanto, Philips, and Shell, to name a few of

them -- are wishing them all the best.

We can study the basics of glue through examining one typical

candidate. Let's examine one well-known superstar of modern

adhesion: that wondrous and well-nigh legendary substance known

as "superglue." Superglue, which also travels under the aliases of

SuperBonder, Permabond, Pronto, Black Max, Alpha Ace, Krazy Glue

and (in Mexico) Kola Loka, is known to chemists as cyanoacrylate

(C5H5NO2).

Cyanoacrylate was first discovered in 1942 in a search for

materials to make clear plastic gunsights for the second world war.

The American researchers quickly rejected cyanoacrylate because

the wretched stuff stuck to everything and made a horrible mess. In

1951, cyanoacrylate was rediscovered by Eastman Kodak researchers

Harry Coover and Fred Joyner, who ruined a perfectly useful

refractometer with it -- and then recognized its true potential.

Cyanoacrylate became known as Eastman compound #910. Eastman

910 first captured the popular imagination in 1958, when Dr Coover

appeared on the "I've Got a Secret" TV game show and lifted host

Gary Moore off the floor with a single drop of the stuff.

This stunt still makes very good television and cyanoacrylate

now has a yearly commercial market of $325 million.

Cyanoacrylate is an especially lovely and appealing glue,

because it is (relatively) nontoxic, very fast-acting, extremely strong,

needs no other mixer or catalyst, sticks with a gentle touch, and does

not require any fancy industrial gizmos such as ovens, presses, vices,

clamps, or autoclaves. Actually, cyanoacrylate does require a

chemical trigger to cause it to set, but with amazing convenience, that

trigger is the hydroxyl ions in common water. And under natural

atmospheric conditions, a thin layer of water is naturally present on

almost any surface one might want to glue.

Cyanoacrylate is a "thermosetting adhesive," which means that

(unlike sealing wax, pitch, and other "hot melt" adhesives) it cannot

be heated and softened repeatedly. As it cures and sets,

cyanoacrylate becomes permanently crosslinked, forming a tough

and permanent polymer plastic.

In its natural state in its native Superglue tube from the

convenience store, a molecule of cyanoacrylate looks something like

this:

CN

/

CH2=C

\

COOR

The R is a variable (an "alkyl group") which slightly changes

the character of the molecule; cyanoacrylate is commercially

available in ethyl, methyl, isopropyl, allyl, butyl, isobutyl,

methoxyethyl, and ethoxyethyl cyanoacrylate esters. These

chemical variants have slightly different setting properties and

degrees of gooiness.

After setting or "ionic polymerization," however, Superglue

looks something like this:

CN CN CN

| | |

- CH2C -(CH2C)-(CH2C)- (etc. etc. etc)

| | |

COOR COOR COOR

The single cyanoacrylate "monomer" joins up like a series of

plastic popper-beads, becoming a long chain. Within the thickening

liquid glue, these growing chains whip about through Brownian

motion, a process technically known as "reptation," named after the

crawling of snakes. As the reptating molecules thrash, then wriggle,

then finally merely twitch, the once- thin and viscous liquid becomes

a tough mass of fossilized, interpenetrating plastic molecular

spaghetti.

And it is strong. Even pure cyanoacrylate can lift a ton with a

single square-inch bond, and one advanced elastomer-modified '80s

mix, "Black Max" from Loctite Corporation, can go up to 3,100 pounds.

This is enough strength to rip the surface right off most substrates.

Unless it's made of chrome steel, the object you're gluing will likely

give up the ghost well before a properly anchored layer of Superglue

will.

Superglue quickly found industrial uses in automotive trim,

phonograph needle cartridges, video cassettes, transformer

laminations, circuit boards, and sporting goods. But early superglues

had definite drawbacks. The stuff dispersed so easily that it

sometimes precipitated as vapor, forming a white film on surfaces

where it wasn't needed; this is known as "blooming." Though

extremely strong under tension, superglue was not very good at

sudden lateral shocks or "shear forces," which could cause the glue-

bond to snap. Moisture weakened it, especially on metal-to-metal

bonds, and prolonged exposure to heat would cook all the strength

out of it.

The stuff also coagulated inside the tube with annoying speed,

turning into a useless and frustrating plastic lump that no amount of

squeezing of pinpoking could budge -- until the tube burst and and

the thin slippery gush cemented one's fingers, hair, and desk in a

mummified membrane that only acetone could cut.

Today, however, through a quiet process of incremental

improvement, superglue has become more potent and more useful

than ever. Modern superglues are packaged with stabilizers and

thickeners and catalysts and gels, improving heat capacity, reducing

brittleness, improving resistance to damp and acids and alkalis.

Today the wicked stuff is basically getting into everything.

Including people. In Europe, superglue is routinely used in

surgery, actually gluing human flesh and viscera to replace sutures

and hemostats. And Superglue is quite an old hand at attaching fake

fingernails -- a practice that has sometimes had grisly consequences

when the tiny clear superglue bottle is mistaken for a bottle of

eyedrops. (I haven't the heart to detail the consequences of this

mishap, but if you're not squeamish you might try consulting The

Journal of the American Medical Association, May 2, 1990 v263 n17

p2301).

Superglue is potent and almost magical stuff, the champion of

popular glues and, in its own quiet way, something of an historical

advent. There is something pleasantly marvelous, almost Arabian

Nights-like, about a drop of liquid that can lift a ton; and yet one can

buy the stuff anywhere today, and it's cheap. There are many urban

legends about terrible things done with superglue; car-doors locked

forever, parking meters welded into useless lumps, and various tales

of sexual vengeance that are little better than elaborate dirty jokes.

There are also persistent rumors of real-life superglue muggings, in

which victims are attached spreadeagled to cars or plate-glass

windows, while their glue-wielding assailants rifle their pockets at

leisure and then stroll off, leaving the victim helplessly immobilized.

While superglue crime is hard to document, there is no

question about its real-life use for law enforcement. The detection

of fingerprints has been revolutionized with special kits of fuming

ethyl-gel cyanoacrylate. The fumes from a ripped-open foil packet of

chemically smoking superglue will settle and cure on the skin oils

left in human fingerprints, turning the smear into a visible solid

object. Thanks to superglue, the lightest touch on a weapon can

become a lump of plastic guilt, cementing the perpetrator to his

crime in a permanent bond.

And surely it would be simple justice if the world's first

convicted superglue mugger were apprehended in just this way.

"Creation Science"

In the beginning, all geologists and biologists were creationists.

This was only natural. In the early days of the Western scientific

tradition, the Bible was by far the most impressive and potent source

of historical and scientific knowledge.

The very first Book of the Bible, Genesis, directly treated

matters of deep geological import. Genesis presented a detailed

account of God's creation of the natural world, including the sea, the

sky, land, plants, animals and mankind, from utter nothingness.

Genesis also supplied a detailed account of a second event of

enormous import to geologists: a universal Deluge.

Theology was queen of sciences, and geology was one humble

aspect of "natural theology." The investigation of rocks and the

structure of the landscape was a pious act, meant to reveal the full

glory and intricacy of God's design. Many of the foremost geologists

of the 18th and 19th century were theologians: William Buckland,

John Pye Smith, John Fleming, Adam Sedgewick. Charles Darwin

himself was a one-time divinity student.

Eventually the study of rocks and fossils, meant to complement

the Biblical record, began to contradict it. There were published

rumblings of discontent with the Genesis account as early as the

1730s, but real trouble began with the formidable and direct

challenges of Lyell's uniformitarian theory of geology and his disciple

Darwin's evolution theory in biology. The painstaking evidence

heaped in Lyell's *Principles of Geology* and Darwin's *Origin of

Species* caused enormous controversy, but eventually carried the

day in the scientific community.

But convincing the scientific community was far from the end

of the matter. For "creation science," this was only the beginning.

Most Americans today are "creationists" in the strict sense of

that term. Polls indicate that over 90 percent of Americans believe

that the universe exists because God created it. A Gallup poll in

1991 established that a full 47 percent of the American populace

further believes that God directly created humankind, in the present

human form, less than ten thousand years ago.

So "creationism" is not the view of an extremist minority in our

society -- quite the contrary. The real minority are the fewer than

five percent of Americans who are strictly non-creationist. Rejecting

divine intervention entirely leaves one with few solid or comforting

answers, which perhaps accounts for this view's unpopularity.

Science offers no explanation whatever as to why the universe exists.

It would appear that something went bang in a major fashion about

fifteen billion years ago, but the scientific evidence for that -- the

three-degree background radiation, the Hubble constant and so forth

-- does not at all suggest *why* such an event should have happened

in the first place.

One doesn't necessarily have to invoke divine will to explain

the origin of the universe. One might speculate, for instance, that

the reason there is Something instead of Nothing is because "Nothing

is inherently unstable" and Nothingness simply exploded. There's

little scientific evidence to support such a speculation, however, and

few people in our society are that radically anti-theistic. The

commonest view of the origin of the cosmos is "theistic creationism,"

the belief that the Cosmos is the product of a divine supernatural

action at the beginning of time.

The creationist debate, therefore, has not generally been

between strictly natural processes and strictly supernatural ones, but

over *how much* supernaturalism or naturalism one is willing to

admit into one's worldview.

How does one deal successfully with the dissonance between

the word of God and the evidence in the physical world? Or the

struggle, as Stephen Jay Gould puts it, between the Rock of Ages and

the age of rocks?

Let us assume, as a given, that the Bible as we know it today is

divinely inspired and that there are no mistranslations, errors,

ellipses, or deceptions within the text. Let us further assume that

the account in Genesis is entirely factual and not metaphorical, poetic

or mythical.

Genesis says that the universe was created in six days. This

divine process followed a well-defined schedule.

Day 1. God created a dark, formless void of deep waters, then

created light and separated light from darkness.

Day 2. God established the vault of Heaven over the formless watery

void.

Day 3. God created dry land amidst the waters and established

vegetation on the land.

Day 4. God created the sun, the moon, and the stars, and set them

into the vault of heaven.

Day 5. God created the fish of the sea and the fowl of the air.

Day 6. God created the beasts of the earth and created one male and

one female human being.

On Day 7, God rested.

Humanity thus began on the sixth day of creation. Mankind is

one day younger than birds, two days younger than plants, and

slightly younger than mammals. How are we to reconcile this with

scientific evidence suggesting that the earth is over 4 billion years

old and that life started as a single-celled ooze some three billion

years ago?

The first method of reconciliation is known as "gap theory."

The very first verse of Genesis declares that God created the heaven

and the earth, but God did not establish "Day" and "Night" until the

fifth verse. This suggests that there may have been an immense

span of time, perhaps eons, between the creation of matter and life,

and the beginning of the day-night cycle. Perhaps there were

multiple creations and cataclysms during this period, accounting for

the presence of oddities such as trilobites and dinosaurs, before a

standard six-day Edenic "restoration" around 4,000 BC.

"Gap theory" was favored by Biblical scholar Charles Scofield,

prominent '30s barnstorming evangelist Harry Rimmer, and well-

known modern televangelist Jimmy Swaggart, among others.

The second method of reconciliation is "day-age theory." In

this interpretation, the individual "days" of the Bible are considered

not modern twenty-four hour days, but enormous spans of time.

Day-age theorists point out that the sun was not created until Day 4,

more than halfway through the process. It's difficult to understand

how or why the Earth would have a contemporary 24-hour "day"

without a Sun. The Beginning, therefore, likely took place eons ago,

with matter created on the first "day," life emerging on the third

"day," the fossil record forming during the eons of "days" four five

and six. Humanity, however, was created directly by divine fiat and

did not "evolve" from lesser animals.

Perhaps the best-known "day-age" theorist was William

Jennings Bryan, three-times US presidential candidate and a

prominent figure in the Scopes evolution trial in 1925.

In modern creation-science, however, both gap theory and

day-age theory are in eclipse, supplanted and dominated by "flood

geology." The most vigorous and influential creation-scientists

today are flood geologists, and their views (though not the only

views in creationist doctrine), have become synonymous with the

terms "creation science" and "scientific creationism."

"Flood geology" suggests that this planet is somewhere between

6,000 and 15,000 years old. The Earth was entirely lifeless until the

six literal 24-hour days that created Eden and Adam and Eve. Adam

and Eve were the direct ancestors of all human beings. All fossils,

including so-called pre-human fossils, were created about 3,000 BC

during Noah's Flood, which submerged the entire surface of the Earth

and destroyed all air-breathing life that was not in the Ark (with the

possible exception of air-breathing mammalian sea life). Dinosaurs,

which did exist but are probably badly misinterpreted by geologists,

are only slightly older than the human race and were co-existent

with the patriarchs of the Old Testament. Actually, the Biblical

patriarchs were contemporaries with all the creatures in the fossil

record, including trilobites, pterosaurs, giant ferns, nine-foot sea

scorpions, dragonflies two feet across, tyrannosaurs, and so forth.

The world before the Deluge had a very rich ecology.

Modern flood geology creation-science is a stern and radical

school. Its advocates have not hesitated to carry the war to their

theological rivals. The best known creation-science text (among

hundreds) is probably *The Genesis Flood: The Biblical Record and

its Scientific Implications* by John C. Whitcomb and Henry M.

Morris (1961). Much of this book's argumentative energy is devoted

to demolishing gap theory, and especially, the more popular and

therefore more pernicious day-age theory.

Whitcomb and Morris point out with devastating logic that

plants, created on Day Three, could hardly have been expected to

survive for "eons" without any daylight from the Sun, created on Day

Four. Nor could plants pollinate without bees, moths and butterflies

-- winged creatures that were products of Day Five.

Whitcomb and Morris marshal a great deal of internal Biblical

testimony for the everyday, non-metaphorical, entirely real-life

existence of Adam, Eve, Eden, and Noah's Flood. Jesus Christ Himself

refers to the reality of the Flood in Luke 17, and to the reality of

Adam, Eve, and Eden in Matthew 19.

Creationists have pointed out that without Adam, there is no

Fall; with no Fall, there is no Atonement for original sin; without

Atonement, there can be no Savior. To lack faith in the historical

existence and the crucial role of Adam, therefore, is necessarily to

lack faith in the historical existence and the crucial role of Jesus.

Taken on its own terms, this is a difficult piece of reasoning to refute,

and is typical of Creation-Science analysis.

To these creation-scientists, the Bible is very much all of a

piece. To begin pridefully picking and choosing within God's Word

about what one may or may not choose to believe is to risk an utter

collapse of faith that can only result in apostasy -- "going to the

apes." These scholars are utterly and soberly determined to believe

every word of the Bible, and to use their considerable intelligence to

prove that it is the literal truth about our world and our history as a

species.

Cynics might wonder if this activity were some kind of

elaborate joke, or perhaps a wicked attempt by clever men to garner

money and fame at the expense of gullible fundamentalist

supporters. Any serious study of the lives of prominent Creationists

establishes that this is simply not so. Creation scientists are not

poseurs or hypocrites. Many have spent many patient decades in

quite humble circumstances, often enduring public ridicule, yet still

working selflessly and doggedly in the service of their beliefs.

When they state, for instance, that evolution is inspired by Satan and

leads to pornography, homosexuality, and abortion, they are entirely

in earnest. They are describing what they consider to be clear and

evident facts of life.

Creation-science is not standard, orthodox, respectable science.

There is, and always has been, a lot of debate about what qualities an

orthodox and respectable scientific effort should possess. It can be

stated though that science should have at least two basic

requirements: (A) the scientist should be willing to follow the data

where it leads, rather than bending the evidence to fit some

preconceived rationale, and (B) explanations of phenomena should

not depend on unique or nonmaterial factors. It also helps a lot if

one's theories are falsifiable, reproducible by other researchers,

openly published and openly testable, and free of obvious internal

contradictions.

Creation-science does not fit that description at all. Creation-

science considers it sheer boneheaded prejudice to eliminate

miraculous, unique explanations of world events. After all, God, a

living and omnipotent Supreme Being, is perfectly capable of

directing mere human affairs into any direction He might please. To

simply eliminate divine intervention as an explanation for

phenomena, merely in order to suit the intellectual convenience of

mortal human beings, is not only arrogant and arbitrary, but absurd.

Science has accomplished great triumphs through the use of

purely naturalistic explanations. Over many centuries, hundreds of

scientists have realized that some questions can be successfully

investigated using naturalistic techniques. Questions that cannot be

answered in this way are not science, but instead are philosophy, art,

or theology. Scientists assume as a given that we live in a natural

universe that obeys natural laws.

It's conceivable that this assumption might not be the case.

The entire cognitive structure of science hinges on this assumption of

natural law, but it might not actually be true. It's interesting to

imagine the consequences for science if there were to be an obvious,

public, irrefutable violation of natural law.

Imagine that such a violation took place in the realm of

evolutionary biology. Suppose, for instance, that tonight at midnight

Eastern Standard Time every human being on this planet suddenly

had, not ten fingers, but twelve. Suppose that all our children were

henceforth born with twelve fingers also and we now found

ourselves a twelve-fingered species. This bizarre advent would

violate Neo-Darwinian evolution, many laws of human metabolism,

the physical laws of conservation of mass and energy, and quite a

few other such. If such a thing were to actually happen, we would

simply be wrong about the basic nature of our universe. We

thought we were living in a world where evolution occurred through

slow natural processes of genetic drift, mutation, and survival of the

fittest; but we were mistaken. Where the time had come for our

species to evolve to a twelve-fingered status, we simply did it in an

instant all at once, and that was that.

This would be a shock to the scientific worldview equivalent to

the terrible shock that the Christian worldview has sustained

through geology and Darwinism. If a shock of this sort were to strike

the scientific establishment, it would not be surprising to see

scientists clinging, quite irrationally, to their naturalist principles --

despite the fact that genuine supernaturalism was literally right at

hand. Bizarre rationalizations would surely flourish -- queer

"explanations" that the sixth fingers had somehow grown there

naturally without our noticing, or perhaps that the fingers were mere

illusions and we really had only ten after all, or that we had always

had twelve fingers and that all former evidence that we had once

had ten fingers were evil lies spread by wicked people to confuse us.

The only alternative would be to fully face the terrifying fact that a

parochial notion of "reality" had been conclusively toppled, thereby

robbing all meaning from the lives and careers of scientists.

This metaphor may be helpful in understanding why it is that

Whitcomb and Morris's *Genesis Flood* can talk quite soberly about

Noah storing dinosaurs in the Ark. They would have had to be

*young* dinosaurs, of course.... If we assume that one Biblical cubit

equals 17.5 inches, a standard measure, then the Ark had a volume

of 1,396,000 cubic feet, a carrying capacity equal to that of 522

standard railroad stock cars. Plenty of room!

Many other possible objections to the Ark story are met head-

on, in similar meticulous detail. Noah did not have to search the

earth for wombats, pangolins, polar bears and so on; all animals,

including the exotic and distant ones, were brought through divine

instinct to the site of the Ark for Noah's convenience. It seems

plausible that this divine intervention was, in fact, the beginning of

the migratory instinct in the animal kingdom. Similarly, hibernation

may have been created by God at this time, to keep the thousands of

animals quiet inside the Ark and also reduce the need for gigantic

animal larders that would have overtaxed Noah's crew of eight.

Evidence in the Biblical geneologies shows that pre-Deluge

patriarchs lived far longer than those after the Deluge, suggesting a

radical change in climate, and not for the better. Whitcomb and

Morris make the extent of that change clear by establishing that

before the Deluge it never rained. There had been no rainbows

before the Flood -- Genesis states clearly that the rainbow came into

existence as a sign of God's covenant with Noah. If we assume that

normal diffraction of sunlight by water droplets was still working in

pre-Deluge time (as seems reasonable), then this can only mean that

rainfall did not exist before Noah. Instead, the dry earth was

replenished with a kind of ground-hugging mist (Genesis 2:6).

The waters of the Flood came from two sources: the "fountains

of the great deep" and "the windows of heaven." Flood geologists

interpret this to mean that the Flood waters were subterranean and

also present high in the atmosphere. Before they fell to Earth by

divine fiat, the Flood's waters once surrounded the entire planet in a

"vapor canopy." When the time came to destroy his Creation, God

caused the vapor canopy to fall from outer space until the entire

planet was submerged. That water is still here today; the Earth in

Noah's time was not nearly so watery as it is today, and Noah's seas

were probably much shallower than ours. The vapor canopy may

have shielded the Biblical patriarchs from harmful cosmic radiation

that has since reduced human lifespan well below Methuselah's 969

years.

The laws of physics were far different in Eden. The Second

Law of Thermodynamics likely began with Adam's Fall. The Second

Law of Thermodynamics is strong evidence that the entire Universe

has been in decline since Adam's sin. The Second Law of

Thermodynamics may well end with the return of Jesus Christ.

Noah was a markedly heterozygous individual whose genes had

the entire complement of modern racial characteristics. It is a

fallacy to say that human embryos recapitulate our evolution as a

species. The bumps on human embryos are not actually relic gills,

nor is the "tail" on an embryo an actual tail -- it only resembles one.

Creatures cannot evolve to become more complex because this would

violate the Second Law of Thermodynamics. In our corrupt world,

creatures can only degenerate. The sedimentary rock record was

deposited by the Flood and it is all essentially the same age. The

reason the fossil record appears to show a course of evolution is

because the simpler and cruder organisms drowned first, and were

the first to sift out in the layers of rubble and mud.

Related so baldly and directly, flood geology may seem

laughable, but *The Genesis Flood* is not a silly or comic work. It is

five hundred pages long, and is every bit as sober, straightforward

and serious as, say, a college text on mechanical engineering.

*The Genesis Flood* has sold over 200,000 copies and gone

through 29 printings. It is famous all over the world. Today Henry

M. Morris, its co-author, is the head of the world's most influential

creationist body, the Institute for Creation Research in Santee,

California.

It is the business of the I.C.R. to carry out scientific research on

the physical evidence for creation. Members of the I.C.R. are

accredited scientists, with degrees from reputable mainstream

institutions. Dr. Morris himself has a Ph.D. in engineering and has

written a mainstream textbook on hydraulics. The I.C.R.'s monthly

newsletter, *Acts and Facts,* is distributed to over 100,000 people.

The Institute is supported by private donations and by income from

its frequent seminars and numerous well-received publications.

In February 1993, I called the Institute by telephone and had

an interesting chat with its public relations officer, Mr. Bill Hoesch.

Mr. Hoesch told me about two recent I.C.R. efforts in field research.

The first involves an attempt to demonstrate that lava flows at the

top and the bottom of Arizona's Grand Canyon yield incongruent

ages. If this were proved factual, it would strongly imply that the

thousands of layers of sedimentary rock in this world-famous mile-

deep canyon were in fact all deposited at the same time and that

conventional radiometric methods are, to say the least, gravely

flawed. A second I.C.R. effort should demonstrate that certain ice-

cores from Greenland, which purport to show 160 thousand years of

undisturbed annual snow layers, are in fact only two thousand years

old and have been misinterpreted by mainstream scientists.

Mr. Hoesch expressed some amazement that his Institute's

efforts are poorly and privately funded, while mainstream geologists

and biologists often receive comparatively enormous federal funding.

In his opinion, if the Institute for Creation Research were to receive

equivalent funding with their rivals in uniformitarian and

evolutionary so-called science, then creation-scientists would soon be

making valuable contributions to the nation's research effort.

Other creation scientists have held that the search for oil, gas,

and mineral deposits has been confounded for years by mistaken

scientific orthodoxies. They have suggested that successful flood-

geology study would revolutionize our search for mineral resources

of all kinds.

Orthodox scientists are blinded by their naturalistic prejudices.

Carl Sagan, whom Mr. Hoesch described as a "great hypocrite," is a

case in point. Carl Sagan is helping to carry out a well-funded

search for extraterrestrial life in outer space, despite the fact that

there is no scientific evidence whatsoever for extraterrestrial

intelligence, and there is certainly no mention in the Bible of any

rival covenant with another intelligent species. Worse yet, Sagan

boasts that he could detect an ordered, intelligent signal from space

from the noise and static of mere cosmic debris. But here on earth

we have the massively ordered and intelligently designed "signal"

called DNA, and yet Sagan publicly pretends that DNA is the result of

random processes! If Sagan used the same criteria to distinguish

intelligence from chance in the study of Earth life, as he does in his

search for extraterrestrial life, then he would have to become a

Creationist!

I asked Mr Hoesch what he considered the single most

important argument that his group had to make about scientific

creationism.

"Creation versus evolution is not science versus religion," he

told me. "It's the science of one religion versus the science of

another religion."

The first religion is Christianity; the second, the so-called

religion of Secular Humanism. Creation scientists consider this

message the single most important point they can make; far more

important than so-called physical evidence or the so-called scientific

facts. Creation scientists consider themselves soldiers and moral

entrepreneurs in a battle of world-views. It is no accident, to their

mind, that American schools teach "scientific" doctrines that are

inimical to fundamentalist, Bible-centered Christianity. It is not a

question of value-neutral facts that all citizens in our society should

quietly accept; it is a question of good versus evil, of faith versus

nihilism, of decency versus animal self-indulgence, and of discipline

versus anarchy. Evolution degrades human beings from immortal

souls created in God's Image to bipedal mammals of no more moral

consequence than other apes. People who do not properly value

themselves or others will soon lose their dignity, and then their

freedom.

Science education, for its part, degrades the American school

system from a localized, community-responsible, democratic

institution teaching community values, to an amoral indoctrination-

machine run by remote and uncaring elitist mandarins from Big

Government and Big Science.

Most people in America today are creationists of a sort. Most

people in America today care little if at all about the issue of creation

and evolution. Most people don't really care much if the world is six

billion years old, or six thousand years old, because it doesn't

impinge on their daily lives. Even radical creation-scientists have

done very little to combat the teaching of evolution in higher

education -- university level or above. They are willing to let Big

Science entertain its own arcane nonsense -- as long as they and

their children are left in peace.

But when world-views collide directly, there is no peace. The

first genuine counter-attack against evolution came in the 1920s,

when high-school education suddenly became far more widely

spread. Christian parents were shocked to hear their children

openly contradicting God's Word and they felt they were losing

control of the values taught their youth. Many state legislatures in

the USA outlawed the teaching of evolution in the 1920s.

In 1925, a Dayton, Tennessee high school teacher named John

Scopes deliberately disobeyed the law and taught evolution to his

science class. Scopes was accused of a crime and tried for it, and his

case became a national cause celebre. Many people think the

"Scopes Monkey Trial" was a triumph for science education, and it

was a moral victory in a sense, for the pro-evolution side

successfully made their opponents into objects of national ridicule.

Scopes was found guilty, however, and fined. The teaching of

evolution was soft-pedalled in high-school biology and geology texts

for decades thereafter.

A second resurgence of creationist sentiment took place in the

1960s, when the advent of Sputnik forced a reassessment of

American science education. Fearful of falling behind the Soviets in

science and technology, the federal National Science Foundation

commissioned the production of state-of-the-art biology texts in

1963. These texts were fiercely resisted by local religious groups

who considered them tantamount to state-supported promotion of

atheism.

The early 1980s saw a change of tactics as fundamentalist

activists sought equal time in the classroom for creation-science -- in

other words, a formal acknowledgement from the government that

their world-view was as legitimate as that of "secular humanism."

Fierce legal struggles in 1982, 1985 and 1987 saw the defeat of this

tactic in state courts and the Supreme Court.

This legal defeat has by no means put an end to creation-

science. Creation advocates have merely gone underground, no

longer challenging the scientific authorities directly on their own

ground, or the legal ground of the courts, but concentrating on grass-

roots organization. Creation scientists find their messages received

with attention and gratitude all over the Christian world.

Creation-science may seem bizarre, but it is no more irrational

than many other brands of cult archeology that find ready adherents

everywhere. All over the USA, people believe in ancient astronauts,

the lost continents of Mu, Lemuria or Atlantis, the shroud of Turin,

the curse of King Tut. They believe in pyramid power, Velikovskian

catastrophism, psychic archeology, and dowsing for relics. They

believe that America was the cradle of the human race, and that

PreColumbian America was visited by Celts, Phoenicians, Egyptians,

Romans, and various lost tribes of Israel. In the high-tech 1990s, in

the midst of headlong scientific advance, people believe in all sorts of

odd things. People believe in crystals and telepathy and astrology

and reincarnation, in ouija boards and the evil eye and UFOs.

People don't believe these things because they are reasonable.

They believe them because these beliefs make them feel better.

They believe them because they are sick of believing in conventional

modernism with its vast corporate institutions, its secularism, its

ruthless consumerism and its unrelenting reliance on the cold

intelligence of technical expertise and instrumental rationality.

They believe these odd things because they don't trust what they are

told by their society's authority figures. They don't believe that

what is happening to our society is good for them, or in their

interests as human beings.

The clash of world views inherent in creation-science has

mostly taken place in the United States. It has been an ugly clash in

some ways, but it has rarely been violent. Western society has had a

hundred and forty years to get used to Darwin. Many of the

sternest opponents of creation-science have in fact been orthodox

American Christian theologians and church officials, wary of a

breakdown in traditional American relations of church and state.

It may be that the most determined backlash will come not

from Christian fundamentalists, but from the legions of other

fundamentalist movements now rising like deep-rooted mushrooms

around the planet: from Moslem radicals both Sunni and Shi'ite, from

Hindu groups like Vedic Truth and Hindu Nation, from militant

Sikhs, militant Theravada Buddhists, or from a formerly communist

world eager to embrace half-forgotten orthodoxies. What loyalty do

these people owe to the methods of trained investigation that made

the West powerful and rich?

Scientists believe in rationality and objectivity -- even though

rationality and objectivity are far from common human attributes,

and no human being practices these qualities flawlessly. As it

happens, the scientific enterprise in Western society currently serves

the political and economic interests of scientists as human beings.

As a social group in Western society, scientists have successfully

identified themselves with the practice of rational and objective

inquiry, but this situation need not go on indefinitely. How would

scientists themselves react if their admiration for reason came into

direct conflict with their human institutions, human community, and

human interests?

One wonders how scientists would react if truly rational, truly

objective, truly nonhuman Artificial Intelligences were winning all

the tenure, all the federal grants, and all the Nobels. Suppose that

scientists suddenly found themselves robbed of cultural authority,

their halting efforts to understand made the object of public ridicule

in comparison to the sublime efforts of a new power group --

superbly rational computers. Would the qualities of objectivity and

rationality still receive such acclaim from scientists? Perhaps we

would suddenly hear a great deal from scientists about the

transcendant values of intuition, inspiration, spiritual understanding

and deep human compassion. We might see scientists organizing to

assure that the Pursuit of Truth should slow down enough for them

to keep up. We might perhaps see scientists struggling with mixed

success to keep Artificial Intelligence out of the schoolrooms. We

might see scientists stricken with fear that their own children were

becoming strangers to them, losing all morality and humanity as they

transferred their tender young brains into cool new racks of silicon

ultra-rationality -- all in the name of progress.

But this isn't science. This is only bizarre speculation.

For Further Reading:

THE CREATIONISTS by Ronald L. Numbers (Alfred A. Knopf, 1992).

Sympathetic but unsparing history of Creationism as movement and

doctrine.

THE GENESIS FLOOD: The Biblical Record and its Scientific

Implications by John C. Whitcomb and Henry M. Morris (Presbyterian

and Reformed Publishing Company, 1961). Best-known and most

often-cited Creationist text.

MANY INFALLIBLE PROOFS: Practical and Useful Evidences of

Christianity by Henry M. Morris (CLP Publishers, 1974). Dr Morris

goes beyond flood geology to offer evidence for Christ's virgin birth,

the physical transmutation of Lot's wife into a pillar of salt, etc.

CATALOG of the Institute for Creation Research (P O Box 2667, El

Cajon, CA 92021). Free catalog listing dozens of Creationist

publications.

CULT ARCHAEOLOGY AND CREATIONISM: Understanding

Pseudoscientific Beliefs About the Past edited by Francis B. Harrold

and Raymond A. Eve (University of Iowa Press, 1987). Indignant

social scientists tie into highly nonconventional beliefs about the

past.

"Robotica '93"

We are now seven years away from the twenty-first

century. Where are all our robots?

A faithful reader of SF from the 1940s and '50s might

be surprised to learn that we're not hip-deep in robots by

now. By this time, robots ought to be making our

breakfasts, fetching our newspapers, and driving our

atomic-powered personal helicopters. But this has not

come to pass, and the reason is simple.

We don't have any robot brains.

The challenge of independent movement and real-time

perception in a natural environment has simply proved too

daunting for robot technology. We can build pieces of

robots in plenty. We have thousands of robot arms in

1993. We have workable robot wheels and even a few

workable robot legs. We have workable sensors for robots

and plenty of popular, industrial, academic and military

interest in robotics. But a workable robot brain remains

beyond us.

For decades, the core of artificial-intelligence

research has involved programming machines to build

elaborate symbolic representations of the world. Those

symbols are then manipulated, in the hope that this will

lead to a mechanical comprehension of reality that can

match the performance of organic brains.

Success here has been very limited. In the glorious

early days of AI research, it seemed likely that if a

machine could be taught to play chess at grandmaster

level, then a "simple" task like making breakfast would be

a snap. Alas, we now know that advanced reasoning skills

have very little to do with everyday achievements such as

walking, seeing, touching and listening. If humans had

to "reason out" the process of getting up and walking out

the front door through subroutines and logical deduction,

then we'd never budge from the couch. These are things

we humans do "automatically," but that doesn't make them

easy -- they only seem easy to us because we're organic.

For a robot, "advanced" achievements of the human brain,

such as logic and mathematical skill, are relatively easy

to mimic. But skills that even a mouse can manage

brilliantly are daunting in the extreme for machines.

In 1993, we have thousands of machines that we

commonly call "robots." We have robot manufacturing

companies and national and international robot trade

associations. But in all honesty, those robots of 1993

scarcely deserve the name. The term "robot" was invented

in 1921 by the Czech playwright Karel Capek, for a stage

drama. The word "robot" came from the Czech term for

"drudge" or "serf." Capek's imaginary robots were made

of manufactured artificial flesh, not metal, and were very

humanlike, so much so that they could actually have sex

and reproduce (after exterminating the humans that created

them). Capek's "robots" would probably be called

"androids" today, but they established the general concept

for robots: a humanoid machine.

If you look up the term "robot" in a modern

dictionary, you'll find that "robots" are supposed to be

machines that resemble human beings and do mechanical,

routine tasks in response to commands.

Robots of this classic sort are vanishingly scarce in

1993. We simply don't have any proper brains for them,

and they can scarcely venture far off the drawing board

without falling all over themselves. We do, however, have

enormous numbers of mechanical robot arms in daily use

today. The robot industry in 1993 is mostly in the

business of retailing robot arms.

There's a rather narrow range in modern industry for

robot arms. The commercial niche for robotics is menaced

by cheap human manual labor on one side and by so-called

"hard automation" on the other. This niche may be

narrow, but it's nevertheless very real; in the US alone,

it's worth about 500 million dollars a year. Over the

past thirty years, a lot of useful technological lessons

have been learned in the iron-arms industry.

Japan today possesses over sixty percent of the entire

world population in robots. Japanese industry won this

success by successfully ignoring much of the glamorized

rhetoric of classic robots and concentrating on actual

workaday industrial uses for a brainless robot arm.

European and American manufacturers, by contrast, built

overly complex, multi-purpose, sophisticated arms with

advanced controllers and reams of high-level programming

code. As a result, their reliability was poor, and in the

grueling environment of the assembly line, they frequently

broke down. Japanese robots were less like the SF concept

of robots, and therefore flourished rather better in the

real world. The simpler Japanese robots were highly

reliable, low in cost, and quick to repay their

investment.

Although Americans own many of the basic patents in

robotics, today there are no major American robot

manufacturers. American robotics concentrates on narrow,

ultra-high-tech, specialized applications and, of course,

military applications. The robot population in the

United States in 1992 was about 40,000, most of them in

automobile manufacturing. Japan by contrast has a

whopping 275,000 robots (more or less, depending on the

definition). Every First World economy has at least some

machines they can proudly call robots; Germany about

30,000, Italy 9,000 or so, France around 13,000, Britain

8,000 and so forth. Surprisingly, there are large numbers

in Poland and China.

Robot arms have not grown much smarter over the years.

Making them smarter has so far proved to be commercially

counterproductive. Instead, robot arms have become much

better at their primary abilities: repetition and

accuracy. Repetition and accuracy are the real selling-

points in the robot arm biz. A robot arm was once

considered a thing of loveliness if it could reliably

shove products around to within a tenth of an inch or so.

Today, however, robots have moved into microchip assembly,

and many are now fantastically accurate. IBM's "fine

positioner," for instance, has a gripper that floats on a

thin layer of compressed air and moves in response to

computer-controlled electromagnetic fields. It has an

accuracy of two tenths of a micron. One micron is one

millionth of a meter. On this scale, grains of dust loom

like monstrous boulders.

CBW Automation's T-190 model arm is not only accurate,

but wickedly fast. This arm plucks castings from hot

molds in less than a tenth of a second, repeatedly

whipping the products back and forth from 0 to 30 miles

per hour in half the time it takes to blink.

Despite these impressive achievements, however, most

conventional robot arms in 1993 have very pronounced

limits. Few robot arms can move a load heavier than 10

kilograms without severe problems in accuracy. The links

and joints within the arm flex in ways difficult to

predict, especially as wear begins to mount. Of course

it's possible to stiffen the arm with reinforcements, but

then the arm itself becomes ungainly and full of

unpredictable inertia. Worse yet, the energy required to

move a heavier arm adds to manufacturing costs. Thanks to

this surprising flimsiness in a machine's metal arm, the

major applications for industrial robots today are

welding, spraying, coating, sealing, and gluing. These

are activities that involve a light and steady movement of

relatively small amounts of material.

Robots thrive in the conditions known in the industry

as "The 3 D's": Dirty, Dull, and Dangerous. If it's too

hot, too cold, too dark, too cramped, or, best of all, if

it's toxic and/or smells really bad, then a robot may well

be just your man for the job!

When it comes to Dirty, Dull and Dangerous, few groups

in the world can rival the military. It's natural

therefore that military-industrial companies such as

Grumman, Martin Marietta and Westinghouse are extensively

involved in modern military-robotics. Robot weaponry and

robot surveillance fit in well with modern US military

tactical theory, which emphasizes "force multipliers" to

reduce US combat casualties and offset the relative US

weakness in raw manpower.

In a recent US military wargame, the Blue or Friendly

commander was allowed to fortify his position with

experimental smart mines, unmanned surveillance planes,

and remote-controlled unmanned weapons platforms. The Red

or Threat commander adamantly refused to take heavy

casualties by having his men battle mere machinery.

Instead, the Threat soldiers tried clumsily to maneuver

far around the flanks so as to engage the human soldiers

in the Blue Force. In response, though, the Blue

commander simply turned off the robots and charged into

the disordered Red force, clobbering them.

This demonstrates that "dumb machines" needn't be

very smart at all to be of real military advantage. They

don't even necessarily have to be used in battle -- the

psychological advantage alone is very great. The US

military benefits enormously if can exchange the potential

loss of mere machinery for suffering and damaged morale in

the human enemy.

Among the major robotics initiatives in the US arsenal

today are Navy mine-detecting robots, autonomous

surveillance aircraft, autonomous surface boats, and

remotely-piloted "humvee" land vehicles that can carry and

use heavy weaponry. American tank commanders are

especially enthused about this idea, especially for

lethally dangerous positions like point-tank in assaults

on fortified positions.

None of these military "robots" look at all like a

human being. They don't have to look human, and in fact

work much better if they don't. And they're certainly not

programmed to obey Asimov's Three Laws of Robotics. If

they had enough of a "positronic brain" to respect the

lives of their human masters, then they'd be useless.

Recently there's been a remarkable innovation in the

"no-brain" approach to robotics. This is the robotic bug.

Insects have been able to master many profound abilities

that frustrate even the "smartest" artificial

intelligences. MIT's famous Insect Lab is a world leader

in this research, building tiny and exceedingly "stupid"

robots that can actually rove and scamper about in rough

terrain with impressively un-robot-like ease.

These bug robots are basically driven by simple

programs of "knee-jerk reflexes." Robot bugs have no

centralized intelligence and no high-level programming.

Instead, they have a decentralized network of simple

abilities that are only loosely coordinated. These

robugs have no complex internal models, and no

comprehensive artificial "understanding" of their

environment. They're certainly not human-looking, and

they can't follow spoken orders. It's been suggested

though that robot bugs might be of considerable commercial

use, perhaps cleaning windows, scavenging garbage, or

repeatedly vacuuming random tiny paths through the carpet

until they'd cleaned the whole house.

If you owned robot bugs, you'd likely never see them.

They'd come with the house, just like roaches or termites,

and they'd emerge only at night. But instead of rotting

your foundation and carrying disease, they'd modestly tidy

up for you.

Today robot bugs are being marketed by IS Robotics of

Cambridge, MA, which is selling them for research and also

developing a home robotic vacuum cleaner.

A swarm of bugs is a strange and seemingly rather

far-fetched version of the classic "household robot." But

the bug actually seems rather more promising than the

standard household robot in 1993, such as the Samsung

"Scout-About." This dome-topped creation, which weighs 16

lbs and is less than a foot high, is basically a mobile

home-security system. It rambles about the house on its

limited battery power, sensing for body-heat, sudden

motion, smoke, or the sound of breaking glass. Should

anything untoward occur, Scout-About calls the police

and/or sets off alarms. It costs about a thousand

dollars. Sales of home-security robots have been less

than stellar. It appears that most people with a need for

such a device would still rather get themselves a dog.

There is an alternative to the no-brain approach in

contemporary robotics. That's to use the brain of a human

being, remotely piloting a robot body. The robot then

becomes "the tele-operated device." Tele-operated robots

face much the same series of career opportunities as their

brainless cousins -- Dirty, Dull and Dangerous. In this

case, though, the robot may be able to perform some of the

Dull parts on its own, while the human pilot successfully

avoids the Dirt and Danger. Many applications for

military robotics are basically tele-operation, where a

machine can maintain itself in the field but is piloted by

human soldiers during important encounters. Much the same

goes for undersea robotics, which, though not a thriving

field, does have niches in exploration, oceanography,

underwater drilling-platform repair, and underwater cable

inspection. The wreck of the *Titanic* was discovered and

explored through such a device.

One of the most interesting new applications of tele-

operated robotics is in surgical tele-operations.

Surgery is, of course, a notoriously delicate and

difficult craft. It calls for the best dexterity humans

can manage -- and then some. A table-mounted iron arm can

be of great use in surgery, because of its swiftness and

its microscopic precision. Unlike human surgeons, a

robot arm can grip an instrument and hold it in place for

hours, then move it again swiftly at a moment's notice

without the least tremor. Robot arms today, such as the

ROBODOC Surgical Assistant System, are seeing use in hip

replacement surgery.

Often the tele-operated robot's grippers are tiny and

at the end of a long flexible cable. The "laparoscope" is

a surgical cable with a tiny light, camera and cutters at

one end. It's inserted through a small hole in the

patient's abdominal wall. The use of laparoscopes is

becoming common, since their use much reduces the shock

and trauma of major surgery.

"Laparoscopy" usually requires two human surgeons,

though; one to cut, and one to guide the cable and camera.

There are obvious potential problems here from missed

communications or simple human exhaustion. With Britain's

"Laparobot," however, a single surgeon can control the

camera angle through a radio-transmitting headband. If he

turns his head, the laparoscope camera pans; if he raises

or lowers his head it tilts up and down, and if he leans

in, then it zooms. And he still has his hands free to

control the blades. The Laparobot is scheduled for

commercial production in late 1993.

Tele-operation has made remarkable advances recently

with the advent of fiber-optics and high-speed computer

networking. However, tele-operation still has very little

to do with the classic idea of a human-shaped robot that

can understand and follow orders. Periodically, there are

attempts to fit the human tele-operator into a human-

shaped remote shell -- something with eyes and arms,

something more traditionally robotlike. And yet, the

market for such a machine has never really materialized.

Even the military, normally not disturbed by commercial

necessity, has never made this idea work (though not from

lack of trying).

The sensory abilities of robots are still very

primitive. Human hands have no less than twenty different

kinds of nerve fiber. Eight kinds of nerve control

muscles, blood vessels and sweat-glands, while the other

twelve kinds sense aspects of pain, temperature, texture,

muscle condition and the angles of knuckles and joints.

No remote-controlled robot hand begins to match this

delicate and sophisticated sensory input.

If robot hands this good existed, they would obviously

do very well as medical prosthetics. It's still

questionable whether there would be a real-world use and

real-world market for a remotely-controlled tele-operated

humanlike robot. There are many industrial uses for

certain separate aspects of humanity -- our grip, our

vision, our propensity for violence -- but few for a

mechanical device with the actual shape and proportions of

a human being.

It seems that our fascination with humanoid robots has

little to do with industry, and everything to do with

society. Robots are appealing for social reasons.

Robots are romantic and striking. Robots have good image.

Even "practical" industrial robots, mere iron arms,

have overreached themselves badly in many would-be

applications. There have been waves of popular interest

and massive investment in robotics, but even during its

boom years, the robot industry has not been very

profitable. In the mid-1980s there were some 300 robot

manufacturers; today there are less than a hundred. In

many cases, robot manufacturers survive because of

deliberate government subsidy. For a nation to own robots

is like owning rocketships or cyclotrons; robots are a

symbol of national technological prowess. Robots mark a

nation as possessing advanced First World status.

Robots are prestige items. In Japan, robots can

symbolize the competition among Japanese firms. This is

why Japanese companies sometimes invent oddities such as

"Monsieur," a robot less than a centimeter across, or a

Japanese boardroom robot that can replace chairs after a

meeting. (Of course one can find human office help to

replace chairs at very little cost and with great

efficiency. But the Japanese office robot replaces

chairs with an accuracy of millimeters!)

It makes a certain sense to subsidize robots. Robots

support advanced infrastructure through their demand-pull

in electronics, software, sensor technology, materials

science, and precision engineering. Spin-offs from

robotics can vitalize an economy, even if the robots

themselves turn out to be mostly decorative. Anyway, if

worst comes to worst, robots have always made excellent

photo-op backgrounds for politicians.

Robots truly thrive as entertainers. This is where

robots began -- on the stage, in Mr. Capek's play in

1921. The best-known contemporary robot entertainers are

probably "Crow" and "Tom Servo" from the cable television

show MYSTERY SCIENCE THEATER 3000. These wisecracking

characters who lampoon bad SF films are not "real robots,"

but only puppets in hardshelled drag; but Crow and Tom are

actors, and actors should be forgiven a little pretense.

Disney "animatronic" robots have a long history and still

have a strong appeal. Lately, robot dinosaurs, robot

prehistoric mammals, and robot giant insects have proved

to be enormous crowd-draws, scaring the bejeezus out of

small children (and, if truth be told, their parents).

Mark Pauline's "Survival Research Laboratories" has won an

international reputation for its violent and catastrophic

robot performance-art. In Austin Texas, the Robot Group

has won a city arts grant to support its robot blimps and

pneumatically-controlled junk-creations.

Man-shaped robots are romantic. They have become

symbols of an early attitude toward technology which, in a

more suspicious and cynical age, still has its own charm

and appeal. In 1993, "robot nostalgia" has become a

fascinating example of how high-tech dreams of the future

can, by missing their target, define their own social

period. Today, fabulous prices are paid at international

antique toy collections for children's toy robots from the

'40s and '50s. These whirring, blinking creatures with

their lithographed tin and folded metal tabs exert a

powerful aesthetic pull on their fanciers. A mint-in-

the-box Robby Robot from 1956, complete with his Space

Patrol Moon Car, can bring over four thousand dollars at

an auction at Christie's. Thunder Robot, a wondrous

creation with machine-gun arms, flashing green eyes, and

whirling helicopter blades over its head, is worth a

whopping nine grand.

Perhaps we like robots better in 1993 because we can't

have them in real life. In today's world, any robot

politely and unquestioningly "obeying human orders" in

accord with Asimov's Three Laws of Robotics would face

severe difficulties. If it were worth even half of what

the painted-tin Thunder Robot is worth, then a robot

streetsweeper, doorman or nanny would probably be beaten

sensorless and carjacked by a gang of young human

unemployables. It's a long way back to yesterday's

tomorrows.

"Watching the Clouds"

In the simmering depths of a Texas summer, there are

few things more soothing than sprawling on a hillside and

watching the clouds roll by. Summer clouds are

especially bright and impressive in Texas, for reasons we

will soon come to understand-- and anyhow, during a Texas

summer, any activity more strenuous than lying down,

staring at clouds, and chewing a grass-stem may well cause

heat-stroke.

By the early nineteenth century, the infant science of

meteorology had freed itself from the ancient Aristotelian

dogma of vapors, humors, and essences. It was known that

the atmosphere was made up of several different gases.

The behavior of gases in changing conditions of heat,

pressure and density was fairly well understood.

Lightning was known to be electricity, and while

electricity itself remained enormously mysterious, it was

under intense study. Basic weather instruments -- the

thermometer, barometer, rain gauge, and weathervane --

were becoming ever more accurate, and were increasingly

cheap and available.

And, perhaps most importantly, a network of amateur

natural philosophers were watching the clouds, and

systematically using instruments to record the weather.

Farmers and sailors owed their lives and livelihoods

to their close study of the sky, but their understanding

was folkloric, not basic. Their rules of thumb were

codified in hundreds of folk weather-proverbs. "When

clouds appear like rocks and towers/ the earth's refreshed

with frequent showers." "Mackerel skies and mares'

tails/ make tall ships carry low sails." This beats

drowning at sea, but it can't be called a scientific

understanding.

Things changed with the advent of Luke Howard, "the

father of British meteorology." Luke Howard was not a

farmer or sailor -- he was a Quaker chemist. Luke Howard

was born in metropolitan London in 1772, and he seems to

have spent most of his life indoors in the big city,

conducting the everyday business of his chemist's shop.

Luke Howard wasn't blessed with high birth or a formal

education, but he was a man of lively and inquiring mind.

While he respected folk weather-wisdom, he also regarded

it, correctly, as "a confused mass of simple aphorisms."

He made it his life's avocation to set that confusion

straight.

Luke Howard belonged to a scientific amateur's club in

London known as the Askesian Society. It was thanks to

these amateur interests that Howard became acquainted with

the Linnaean System. Linnaeus, an eighteenth-century

Swedish botanist, had systematically ranked and classified

the plants and animals, using the international language

of scholarship, Latin. This highly useful act of

classification and organization was known as

"modification" in the scientific terminology of the time.

Though millions of people had watched, admired, and

feared clouds for tens of thousands of years, it was Luke

Howard's particular stroke of genius to recognize that

clouds might also be classified.

In 1803, the thirty-one-year-old Luke Howard presented

a learned paper to his fellow Askesians, entitled "On the

Modifications of Clouds, and On the Principles of Their

Production, Suspension, and Destruction."

Howard's speculative "principles" have not stood the

test of time. Like many intellectuals of his period,

Howard was utterly fascinated by "electrical fluid," and

considered many cloud shapes to be due to static

electricity. Howard's understanding of thermodynamics was

similarly halting, since, like his contemporaries, he

believed heat to be an elastic fluid called Caloric.

However, Howard's "modifications" -- cirrus, cumulus,

and stratus -- have lasted very successfully to the

present day and are part of the bedrock of modern

meteorology. Howard's scholarly reputation was made by

his "modifications," and he was eventually invited to join

the prestigious Royal Society. Luke Howard became an

author, lecturer, editor, and meteorological instrument-

maker, and a learned correspondent with superstars of

nineteenth-century scholarship such as Dalton and Goethe.

Luke Howard became the world's recognized master of

clouds. In order to go on earning a living, though, the

father of British meteorology wisely remained a chemist.

Thanks to Linnaeus and his disciple Howard, cloud

language abounds in elegant Latin constructions. The

"genera" of clouds are cirrus, cirrocumulus, cirrostratus;

altocumulus, altostratus, nimbostratus; stratocumulus,

cumulus and cumulonimbus.

Clouds can also be classified into "species," by their

peculiarities in shape and internal structure. A glance

through the World Meteorological Organization's official

*International Cloud Atlas* reveals clouds called:

fibratus, uncinus, spissatus, castellanus, floccus,

stratiformus, nebulosus, lenticularis, fractus, humilis,

mediocris, congestus, calvus, and capillatus.

As if that weren't enough, clouds can be further

divvied-up into "varieties," by their "special

characteristics of arrangement and transparency":

intortus, vertebratus, undulatus, radiatus, lacunosis,

duplicatus, translucidus, perlucidus and opacis.

And, as a final scholastic fillip, there are the nine

supplementary features and appended minor cloud forms:

incus, mammatus, virga, praecipitatio, arcus, tuba,pileus,

vella, and pannus.

Luke Howard had quite a gift for precise language, and

sternly defended his use of scholar's Latin to other

amateurs who would have preferred plain English. However

elegant his terms, though, Howard's primary insight was

simple. He recognized that most clouds come in two basic

types: "cumulus" and "stratus," or heaps and layers.

Heaps are commoner than layers. Heaps are created by

local rising air, while layers tend to sprawl flatly

across large areas.

Water vapor is an invisible gas. It's only when the

vapor condenses, and begins to intercept and scatter

sunlight as liquid droplets or solid ice crystals, that we

can see and recognize a "cloud." Great columns and

gushes of invisible vapor continue to enter and leave the

cloud throughout its lifetime, condensing within it and

evaporating at its edges. This is one reason why clouds

are so mutable -- clouds are something like flames,

wicking along from candles we can't see.

Who can see the wind? But even when we can't feel

wind, the air is always in motion. The Earth spins

ponderously beneath its thin skin of atmosphere, dragging

air with it by gravity, and arcing wind across its surface

with powerful Coriolis force. The strength of sunlight

varies between pole and equator, powering gigantic Hadley

Cells that try to equalize the difference. Mountain

ranges heave air upward, and then drop it like bobsleds

down their far slopes. The sunstruck continents simmer

like frying pans, and the tropical seas spawn giant

whirlpools of airborne damp.

Water vapor moves and mixes freely with all of these

planetary surges, just like the atmosphere's other trace

constituents. Water vapor, however, has a unique quality

-- at Earth's temperatures, water can become solid, liquid

or gas. These changes in form can store, or release,

enormous amounts of heat. Clouds can power themselves by

steam.

A Texas summer cumulus cloud is the child of a rising

thermal, from the sun-blistered Texan earth. Heated air

expands. Expanding air becomes buoyant, and rises. If

no overlying layer of stable air stops it from rising, the

invisible thermal will continue to rise, and cool, until

it reaches the condensation level. The condensation level

is what gives cumulus clouds their flat bases -- to Luke

Howard, the condensation level was colorfully known as

"the Vapour Plane." Depending on local heat and humidity,

the condensation level may vary widely in height, but it's

always up there somewhere.

At this point, the cloud's internal steam-engine kicks

in. Billions of vapor molecules begin to cling to the

enormous variety of trash that blesses our atmosphere:

bits of ash and smoke from volcanoes and forest-fires,

floating spores and pollen-grains, chips of sand and dirt

kicked up by wind-gusts, airborne salt from bubbles

bursting in the ocean, meteoric dust sifting down from

space. As the vapor clings to these "condensation

nuclei," it condenses, and liquefies, and it gives off

heat.

This new gush of heat causes the air to expand once

again, and propels it upward in a rising tower, topped by

the trademark cauliflower bubbles of the summer cumulus.

If it's not disturbed by wind, hot dry air will cool

about ten degrees centigrade for every kilometer that it

rises above the earth. This rate of cooling is known to

Luke Howard's modern-day colleagues as the Dry Adiabatic

Lapse Rate. Hot *damp* air, however, cools in the *Wet*

Adiabatic Lapse Rate, only about six degrees per kilometer

of height. This four-degree difference in energy --

caused by the "latent heat" of the wet air -- is known in

storm-chasing circles as "the juice."

When bodies of wet and dry air collide along what is

known as "the dryline," the juice kicks in with a

vengeance, and things can get intense. Every spring, in

the High Plains of Texas and Oklahoma, dry air from the

center of the continent tackles damp surging warm fronts

from the soupy Gulf of Mexico. The sprawling plains that

lie beneath the dryline are aptly known as "Tornado

Alley."

A gram of condensing water-vapor has about 600

calories of latent heat in it. One cubic meter of hot

damp air can carry up to three grams of water vapor.

Three grams may not seem like much, but there are plenty

of cubic meters in a cumulonimbus thunderhead, which tends

to be about ten thousand meters across and can rise eleven

thousand meters into the sky, forming an angry, menacing

anvil hammered flat across the bottom of the stratosphere.

The resulting high winds, savage downbursts, lashing

hail and the occasional city-wrecking tornado can be

wonderfully dramatic and quite often fatal. However, in

terms of the Earth's total heat-budget, these local

cumulonimbus fireworks don't compare in total power to the

gentle but truly vast stratus clouds. Stratus tends to

be the product of air gently rising across great expanses

of the earth, air that is often merely nudged upward, at a

few centimeters per second, over a period of hours. Vast

weather systems can slowly pump up stratus clouds in huge

sheets, layer after layer of flat overcast that sometimes

covers a quarter of North America.

Fog is also a stratus cloud, usually created by warm

air's contact with the cold night earth. Sometimes a

gentle uplift of moving air, oozing up the long slope from

the Great Plains to the foot of the Rockies, can produce

vast blanketing sheets of ground-level stratus fog that

cover entire states.

As it grows older, stratus cloud tends to break up

into dapples or billows. The top of the stratus layer

cools by radiation into space, while the bottom of the

cloud tends to warm by intercepting the radiated heat from

the earth. This gentle radiant heat creates a mild, slow

turbulence that breaks the solid stratus into thousands of

leopard-spots, or with the aid of a little wind, perhaps

into long billows and parallel rolls. Thicker, lowlying

stratus may not break-up enough to show clear sky, but

simply become a dispiriting mass of gloomy gray knobs and

lumps that can last for days on end, during a quiet

winter.

When vapor condenses into droplets, it gives off

latent heat and rises. The cooler air from the heights,

shoved aside by the ascending warm air, tends to fall. If

the falling air drags some captured droplets of water with

it, those droplets will evaporate on the way down. This

makes the downdraft cooler and denser, and speeds its

descent. It's "the juice" again, but in reverse. If

there's enough of this steam-power set-loose, it will

create vertically circulating masses of air, or

"convection cells."

Downdraft winds are invisible, but they are a vital

part of the cloud system. In a patchy summer sky,

downdrafts fill the patches between the clouds --

downdrafts *are* the patches. They tear droplets from

the edges of clouds and consume them.

Most clouds never manage to rain or snow. They simply

use the vapor-water cycle as a mechanism to carry and

dissipate excess heat, doing the Earth's quiet business of

entropy.

Clouds also scour the sky; they are the atmosphere's

cleaning agents. A good rain always makes the air seem

fresh and clean, but even clouds that never rain can

nevertheless clean up billions of dust particles. Tiny

droplets carry their dust nuclei with them as they collide

with one another inside the cloud, and combine into large

drops of water. Even if this drop then evaporates and

never falls as rain, the many dust particles inside it

will congeal thorough adhesion into a good-sized speck,

which will eventually settle to earth on its own.

For a drop of water to fall successfully to earth, it

has to increase in size by about one million times, from

the micron width of a damp condensation nucleus, to the

hefty three millimeters of an honest raindrop. A raindrop

can grow by condensation about to a tenth of a millimeter,

but after this scale is reached, condensation alone will

no longer do the job, and the drop has to rely on

collision and capture.

Warm damp air rising within a typical rainstorm

generally moves upward at about a meter per second.

Drizzle falls about one centimeter per second and so is

carried up with the wind, but as drops grow, their rate of

descent increases. Eventually the larger drops are poised

in midair, struggling to fall, as tiny droplets are swept

up past them and against them. The drop will collide and

fuse with some of the droplets in its path, until it grows

too large for the draft to support. If it is then caught

in a cool downdraft, it may survive to reach the earth as

rain. Sometimes the sheer mass of rain can overpower the

updraft, through accumulating weight and the cooling power

of its own evaporation.

Raindrops can also grow as ice particles at the frigid

tops of tall clouds. "Sublimation" is the process of

water vapor directly changing from water to ice. If the

air is cold enough, ice crystals grow much faster in

saturated air than a water droplet does. An ice crystal

in damp supercooled air can grow to raindrop size in only

ten minutes. An upper-air snowflake, if it melts during

its long descent, falls as rain.

Truly violent updrafts to great heights can create

hail. Violent storms can create updrafts as fast as

thirty meters a second, fast enough to buoy up the kind of

grapefruit-sized hail that sometimes kills livestock and

punches holes right through roofs. Some theorists believe

that the abnormally fat raindrops, often the first signs

of an approaching thundershower, are thin scatterings of

thoroughly molten hail.

Rain is generally fatal to a cumulonimbus cloud,

causing the vital loss of its "juice." The sharp, clear

outlines of its cauliflower top become smudgy and sunken.

The bulges flatten, and the crevasses fill in. If there

are strong winds at the heights, the top of the cloud can

be flattened into an anvil, which, after rain sets in, can

be torn apart into the long fibrous streaks of anvil

cirrus. The lower part of the cloud subsides and

dissolves away with the rain, and the upper part drifts

away with the prevailing wind, slowly evaporating into

broken ragged fragments, "fractocumulus."

However, if there is juice in plenty elsewhere, then a

new storm tower may spring up on the old storm's flank.

Systems of storm will therefore often propagate at an

angle across the prevailing wind, bubbling up to the right

or left edge of an advancing mass of clouds. There may

be a whole line of such storms, bursting into life at one

end, and collapsing into senescence at the other. The

youngest tower, at the far edge of the storm-line, usually

has the advantage of the strongest supply of juice, and is

therefore often the most violent. Storm-chasers tend to

cluster at the storm's trailing edge to keep a wary eye on

"Tail-End Charlie."

Because of the energy it carries, water vapor is the

most influential trace gas in the atmosphere. It's the

only gas in the atmosphere that can vary so drastically,

plentiful at some times and places, vanishing at others.

Water vapor is also the most dramatic gas, because liquid

water, cloud, is the only trace constituent in our

atmosphere that we can actually see.

The air is mostly nitrogen -- about 78 percent.

Oxygen is about 21 percent, argon one percent. The rest

is neon, helium, krypton, hydrogen, xenon, ozone and just

a bit of methane and carbon dioxide. Carbon dioxide,

though vital to plant life, is a vanishingly small 0.03

percent of our atmosphere.

However, thanks to decades of hard work by billions of

intelligent and determined human beings, the carbon

dioxide in our atmosphere has increased by twenty percent

in the last hundred years. During the next fifty years,

the level of carbon dioxide in the atmosphere will

probably double.

It's possible that global society might take coherent

steps to stop this process. But if this process actually

does take place, then we will have about as much chance to

influence the subsequent course of events as the late Luke

Howard.

Carbon dioxide traps heat. Since clouds are our

atmosphere's primary heat-engines, doubling the carbon

dioxide will likely do something remarkably interesting to

our clouds. Despite the best efforts of whirring

supercomputers at global atmospheric models around the

world, nobody really knows what this might be. There are

so many unknown factors in global climatology that our

best speculations on the topic are probably not much more

advanced, comparatively speaking, than the bold but

mistaken theorizing of Luke Howard.

One thing seems pretty likely, though. Whatever our

clouds may do, quite a few of the readers of this column

will be around in fifty years to watch them.

"Spires on the Skyline"

Broadcast towers are perhaps the single most obvious

technological artifact of modern life. At a naive glance,

they seem to exist entirely for their own sake. Nobody

lives in them. There's nothing stored in them, and they

don't offer shelter to anyone or anything. They're

skeletal, forbidding structures that are extremely tall

and look quite dangerous. They stand, usually, on the

highest ground available, so they're pretty hard not to

notice. What's more, they're brightly painted and/or

covered with flashing lights.

And then there are those *things* attached to them.

Antennas of some kind, presumably, but they're nothing

like the normal, everyday receiving antennas you might

have at home: a simple telescoping rod for a radio, a

pair of rabbit ears for a TV. These elaborate,

otherworldly appurtenances resemble big drums, or sea

urchin spines, or antlers.

In this column, we're going to demystify broadcast

towers, and talk about what they do, and why they look

that way, and how they've earned their peculiar right to

loom eerily on the skyline of every urban center in

America.

We begin with the electromagnetic spectrum. Towers

have everything to do with the electromagnetic spectrum.

Basically, they colonize the spectrum. They legally

settle various patches of it, and they use their

homestead in the spectrum to make money for their owners

and users.

The electromagnetic spectrum is an important natural

resource. Unlike most things we think of as "resources,"

the spectrum is immaterial and intangible. Odder still,

it is limited, and yet, it is not exhaustible. Usage of

the spectrum is controlled worldwide by an international

body known as the International Telecommunications Union

(ITU), and controlled within the United States by an

agency called the Federal Communications Commission (FCC).

Electromagnetic radiation comes in a wide variety of

flavors. It's usually discussed in terms of frequency and

wavelength, which are interchangeable terms. All

electromagnetic radiation moves at one uniform speed, the

speed of light. If the frequency of the wave is higher,

then the length of the wave must by necessity become

shorter.

Waves are measured in hertz. One hertz is one cycle

of frequency per second, named after Heinrich Hertz, a

nineteenth-century German physicist who was the first in

history to deliberately send a radio signal.

The International Telecommunications Union determines

the legally possible uses of the spectrum from 9,000 hertz

(9 kilohertz) to 400,000,000,000 hertz (400 gigahertz).

This vast legal domain extends from extremely low

frequency radio waves up to extremely high frequency

microwaves. The behavior of electromagnetic radiation

varies considerably along this great expanse of frequency.

As frequency rises, the reach of the signal deteriorates;

the signal travels less easily, and is more easily

absorbed and scattered by rain, clouds, and foliage.

After electromagnetic radiation leaves the legal

domain of the ITU, its behavior becomes even more

remarkable, as it segues into infrared, then visible

light, then ultraviolet, Xrays, gamma rays and cosmic

rays.

From the point of view of physics, there's a strangely

arbitrary quality to the political decisions of the ITU.

For instance, it would seem very odd if there were an

international regulatory body deciding who could license

and use the color red. Visible colors are a form of

electromagnetism, just like radio and microwaves. "Red"

is a small piece of the electromagnetic spectrum which

happens to be perceivable by the human eye, and yet it

would seem shocking if somebody claimed exclusive use of

that frequency. The spectrum really isn't a "territory"

at all, and can't really be "owned," even though it can

be, and is, literally auctioned off to private bidders by

national governments for very large sums. Politics and

commerce don't matter to the photons. But they matter

plenty to the people who build and use towers.

The ITU holds regular international meetings, the

World Administrative Radio Conferences, in which various

national players jostle over spectrum usage. This is an

odd and little-recognized species of diplomacy, but the

United States takes it with utter seriousness, as do other

countries. The resultant official protocols of global

spectrum usage closely resemble international trade

documents, or maybe income-tax law. They are very arcane,

very specific, and absolutely riddled with archaisms,

loopholes, local exceptions and complex wheeler-dealings

that go back decades. Everybody and his brother has some

toehold in the spectrum: ship navigation, aircraft

navigation, standard time signals, various amateur ham

radio bands, industrial remote-control radio bands, ship-

to-shore telephony, microwave telephone relays, military

and civilian radars, police radio dispatch, radio

astronomy, satellite frequencies, kids' radio-controlled

toys, garage-door openers, and on and on.

The spectrum has been getting steadily more crowded

for decades. Once a broad and lonely frontier, inhabited

mostly by nutty entrepreneurs and kids with crystal sets,

it is now a thriving, uncomfortably crowded metropolis.

In the past twenty years especially, there has been

phenomenal growth in the number of machines spewing radio

and microwave signals into space. New services keep

springing up: telephones in airplanes, wireless

electronic mail, mobile telephones, "personal

communication systems," all of them fiercely demanding

elbow-room.

AM radio, FM radio, and television all have slices of

the spectrum. They stake and hold their claim with

towers. Towers have evolved to fit their specialized

environment: a complex interplay of financial necessity,

the laws of physics, and government regulation.

Towers could easily be a lot bigger than they are.

They're made of sturdy galvanized steel, and the

principles of their construction are well-understood.

Given four million dollars, it would be a fairly simple

matter to build a broadcast tower 4,000 feet high. In

practice, however, you won't see towers much over 2,100

feet in the United States, because the FCC deliberately

stunts them. A broadcast antenna atop a 4000-ft tower

would hog the spectrum over too large a geographical area.

Almost every large urban antenna-tower, the kind you

might see in everyday life, belongs to some commercial

entity. Military and scientific-research antennas are

more discreet, usually located in remote enclaves.

Furthermore, they just don't look like commercial

antennas. Military communication equipment is not

subject to commercial restraints and has a characteristic

appearance: rugged, heavy-duty, clunky, serial-numbered,

basically Soviet-looking. Scientific instruments are

designed to gather data with an accuracy to the last

possible decimal point. They may look frazzled, but they

rarely look simple. Broadcast tower equipment by

contrast is designed to make money, so it looks cheerfully

slimmed-down and mass-produced and gimcrack.

Of course, a commercial antenna must obey the laws of

physics like other antennas, and has been designed to do

that, but its true primary function is generating optimal

revenue on capital investment. Towers and their antennas

cost as little as possible, consonant with optimal

coverage of the market area, and the likelihood of

avoiding federal prosecution for sloppy practices. Modern

antennas are becoming steadily more elaborate, so as to

use thinner slices of spectrum and waste less radiative

power. More elaborate design also reduces the annoyance

of stray, unwanted signals, so-called "electromagnetic

pollution."

Towers fall under the aegis of not one but two

powerful bureaucracies, the FCC and the FAA, or Federal

Aviation Administration. The FAA is enormously fond of

massive air-traffic radar antennas, but dourly regards

broadcast antennas as a "menace to air navigation."

This is the main reason why towers are so flauntingly

obvious. If towers were painted sky-blue they'd be almost

invisible, but they're not allowed this. Towers are

hazards to the skyways, and therefore they are striped in

glaring "aviation white" and gruesome "international

orange," as if they were big traffic sawhorses.

Both the FCC and FAA are big outfits that have been

around quite a while. They may be slow and cumbersome,

but they pretty well know the name of the game. Safety

failures in tower management can draw savage fines of up

to a hundred thousand dollars a day. FCC regional offices

have mandatory tower inspection quotas, and worse yet, the

fines on offenders go tidily right into the FCC's budget.

That orange and white paint costs a lot. It also

peels off every couple of years, and has to be replaced,

by hand. Depending on the size of the tower, it's

sometimes possible to get away with using navigation-

hazard lights instead of paint, especially if the lights

strobe. The size of the lights, and their distribution

on the tower structure, and their wattage, and even their

rate and method of flashing are all spelled out in

grinding detail by the FCC and FAA.

In the real world -- and commercial towers are very

real-world structures -- lights aren't that much of an

advantage over paint. The bulbs burn out, for one thing.

Rain shorts out the line. Ice freezes solid on the high

upper reaches of the tower, plummets off in big thirty-

pound chunks, cracking the lights off (not to mention

cracking the lower-mounted antennas, the hoods and

windshields of utility trucks, and the skulls of unlucky

technicians). The lights' power sometimes fails entirely.

And people shoot the lights and steal them. In the

real world, people shoot towers all the time. Something

about towers -- their dominating size, their lonely

locales, or maybe it's that color-scheme and that pesky

blinking -- seems to provoke an element of trigger-happy

lunacy in certain people. Bullet damage is a major

hassle for the tower owner and renter.

People, especially drunken undergraduates in college

towns, often climb the towers and steal the hazard lights

as trophies. If you visit the base of a tower, you will

usually find it surrounded with eight-foot, padlocked

galvanized fencing and a mean coil of sharp razor-wire.

But that won't stop an active guy with a pickup, a ladder,

and a six-pack under his belt.

The people who physically build and maintain towers

refer to themselves as "tower hands." Tower engineers and

designers refer to these people as "riggers." The suit-

and-tie folks who actually own broadcasting stations refer

to them as "tower monkeys." Tower hands are blue-collar

industrial workers, mostly agile young men, mostly

nonunionized. They're a special breed. Not everybody

can calmly climb 2,000 feet into their air with a twenty-

pound tool-belt of ohmmeters, wattmeters, voltage meters,

and various wrenches, clamps, screwdrivers and specialized

cutting tools. Some people get used to this and come to

enjoy it, but those who don't get used to it, *never* get

used to it.

While 2,000 feet in the air, these unsung knights of

the airwaves must juggle large, unwieldy antennas. Quite

often they work when the station is off the air -- in the

midnight darkness, using helmet-mounted coal-miners'

lamps. And it's hot up there on the tower, or freezing,

or wet, and almost always windy.

The commonest task in the tower-hand's life is

painting. It's done with "paint-mitts," big soppy gloves

dipped in paint, which are stroked over every structural

element in the tower, rather like grooming a horse. It

takes a strong man a full day to paint a hundred feet of

an average tower. (Rip-off hustlers posing as tower-hands

can paint towers at "bargain rates" with amazing

cheapness and speed. The rascals -- there are some in

every business -- paint only the *underside* of the

tower, the parts visible from the ground.)

Spray-on paint can be faster than hand-work, but with

even the least breeze, paint sprayed 2,000 feet up will

carry hundreds of yards to splatter the roofs, walls, and

cars of angry civilians with vivid "international orange."

There simply isn't much calm air 2,000 feet up in the sky.

High-altitude wind doesn't have to deal with ground-level

friction, so wind-speed roughly doubles about every

thousand feet.

Building towers is known in the trade as "stacking

steel." The towers are shipped in pieces, then bolted or

welded into segments, either on-site or at the shop. The

rigid sections are hauled skyward with a winch-driven

'load line,' and kept from swaying by a second steel

cable, the 'tag-line.' Each section is bootstrapped up

above the top of the tower, through the use of a tower-

mounted crane, called the 'gin pole.' The gin pole has a

360-degree revolving device at its very top, the 'rooster

head.' Each new section is deliberately hauled up, spun

deftly around on the rooster head, stacked on top of all

the previous sections, and securely bolted into place.

Then the tower hands detach the gin pole, climb the

section they just stacked, mount the ginpole up at the

top again, and repeat the process till they're done.

Tower construction is a mature industry; there have

not been many innovations in the last forty years.

There's nothing new about galvanized steel; it's not high-

tech, but it's plenty sturdy, it's easy to work and weld,

and it gets the job done. The job's not cheap. In

today's market, galvanized steel towers tend to cost about

a million dollars per thousand feet of height.

Towers come in two basic varieties, self-supporting

and guyed. The self-supporting towers are heavier and

more expensive, their feet broadly splayed across the

earth. Despite their slender spires, the guyed towers

actually require more room. The bottom of a guyed tower

is tapered and quite slender, often a narrow patch of

industrial steel not much bigger than the top of a child's

school-desk. But the foundations for those guy cables

stretch out over a vast area, sometimes 100 percent of the

tower's height, in three or four different directions.

It's possible to draw the cables in toward the tower's

base, but that increases the "download" on the tower

structure.

Towers are generally built as lightly as possible,

commensurate with the strain involved. But the strain

is very considerable. Towers themselves are heavy. They

need to be sturdy enough to have tower-hands climbing any

part of them, at any time, safely.

Small towers sometimes use their bracing bars as

natural step-ladders, but big towers have a further

burden. It takes a strong man, with a clear head, 3/4 of

an hour to climb a thousand feet, so any tower over that

size definitely requires an elevator. That brings the

full elaborate rigging of guide rails, driving mechanism,

hoisting cables, counterweights, rope guards, and cab

controls, all of which add to the weight and strain on the

structure. Even with an elevator, one still needs a

ladder for detail work. Tower hands, who have a very good

head for heights, prefer their ladders out on the open

air, where there are fewer encumbrances, and they can get

the job done in short order. However, station engineers

and station personnel, who sometimes need to whip up the

tower to replace a lightbulb or such, rather prefer a

ladder that's nestled inside the tower, so the structure

itself forms a natural safety cage.

Besides the weight of the tower, its elevator, the

power cables, the waveguides, the lights, and the

antennas, there is also the grave risk of ice. Ice forms

very easily on towers, much like the icing of an aircraft

wing. An ice-storm can add hugely to a tower's weight,

and towers must be designed for that eventuality.

Lightning is another prominent hazard, and although

towers are well-grounded, lightning can be freakish and

often destroys vulnerable antennas and wiring.

But the greatest single threat to a tower is wind-

load. Wind has the advantage of leverage; it can attack

a tower from any direction, anywhere along its length, and

can twist it, bend it, shake it, pound it, and build up

destructive resonant vibrations.

Towers and their antennas are built to avoid resisting

wind. The structural elements are streamlined. Often the

antennas have radomes, plastic weatherproof covers of

various shapes. The plastic radome is transparent to

radio and microwave emissions; it protects the sensitive

antenna and also streamlines it to avoid wind-load.

An antenna is an interface between an electrical

system and the complex surrounding world of moving

electromagnetic fields. Antennas come in a bewildering

variety of shapes, sizes and functions. The Andrew

Corporation, prominent American tower builders and

equipment specialists, sells over six hundred different

models of antennas.

Antennas are classified in four basic varieties:

current elements, travelling-wave antennas, antenna

arrays, and radiating-aperture antennas. Elemental

antennas tend to be low in the frequency range,

travelling-wave antennas rather higher, arrays a bit

higher yet, and aperture antennas deal with high-frequency

microwaves. Antennas are designed to meet certain

performance parameters: frequency, radiation pattern,

gain, impedance, bandwidth, polarization, and noise

temperature.

Elemental antennas are not very "elemental." They

were pretty elemental back in the days of Guglielmo

Marconi, the first to make any money broadcasting, but

Marconi's radiant day of glory was in 1901, and his field

of "Marconi wireless" has enjoyed most of a long century

of brilliant innovation and sustained development.

Monopole antennas are pretty elemental -- just a big metal

rod, spewing out radiation in all directions -- but they

quickly grow more elaborate. There are doublets and

dipoles and loops; slots, stubs, rods, whips; biconal

antennas, spheroidal antennas, microstrip radiators.

Then there's the travelling-wave antennas: rhombic,

slotted waveguides, spirals, helices, slow wave, fast

wave, leaky wave.

And the arrays: broadside, endfire, planar, circular,

multiplicative, beacon, et al.

And aperture variants: the extensive microwave clan.

The reflector family: single, dual, paraboloid,

spherical, cylindrical, off-set, multi-beam, contoured,

hybrid, tracking.... The horn family: pyramidal,

sectoral, conical, biconical, box, hybrid, ridged. The

lens family: metal lens, dielectric lens, Luneberg lens.

Plus backfire aperture, short dielectric rods, and

parabolic horns.

Electromagnetism is a difficult phenomenon. The

behavior of photons doesn't make much horse sense, and is

highly counterintuitive. Even the bedrock of

electromagnetic understanding, Maxwell's equations,

require one to break into specialized notation, and the

integral calculus follows with dreadful speed. To put it

very simply: antennas come in different shapes and sizes

because they are sending signals of different quality, in

fields of different three-dimensional shape.

Wavelength is the most important determinant of

antenna size. Low frequency radiation has a very long

wavelength and works best with a very long antenna. AM

broadcasting is low frequency, and in AM broadcasting the

tower *is* the antenna. The AM tower itself is mounted

on a block of insulation. Power is pumped into the entire

tower and the whole shebang radiates. These low-frequency

radio waves can bounce off the ionosphere and go amazing

distances.

Microwaves, however, are much farther up the spectrum.

Microwave radiation has a short wavelength and behaves

more like light. This is why microwave antennas come as

lenses and dishes, rather like the lens and retina of a

human eye.

An array antenna is a group of antennas which

interreact in complex fashion, bouncing and shaping the

radiation they emit. The upshot is a directional beam.

"Coverage is coverage," as the tower-hands say, so

very often several different companies, or even several

different industries, will share towers, bolting their

equipment up and down the structure, rather like oysters,

limpets and barnacles all settling on the same reef.

Here's a brief naturalist's description of some of the

mechanical organisms one is likely to see on a broadcast

tower.

First -- the largest and most obvious -- are things

that look like big drums. These are microwave dishes

under their protective membranes of radome. They may be

flat on both sides, in which case they are probably two

parabolic dishes mounted back-to-back. They may be flat

on one side, or they may bulge out on both sides so that

they resemble a flying saucer. If they are mounted so

that the dish faces out horizontally, then they are relays

of some kind, perhaps local telephone or a microwave long-

distance service. They might be a microwave television-

feed to a broadcast TV network affiliate, or a local

cable-TV system. They don't broadcast for public

reception, because the microwave beams from these focused

dishes are very narrow. Somewhere in the distance,

probably within 30 miles, is another relay in the chain.

A tower may well have several satellite microwave

dishes. These will be down near the base of the tower,

hooked to the tower by cable and pointed almost straight

up. These satellite dishes are generally much bigger than

relay microwave dishes. They're too big to fit on a

tower, and there's no real reason to put them them on a

tower anyway; they'll scarcely get much closer to an

orbiting satellite by rising a mere 2,000 feet.

Often, small microwave dishes made of metal slats are

mounted to the side of the tower. These slat dishes are

mostly empty space, so they're less electronically

efficient than a smooth metal dish would be. However, a

smooth metal dish, being cupshaped, acts just like the cup

on a wind-gauge, so if a strong wind-gust hits it, it will

strain the tower violently. Slotted dishes are

lighter,cheaper and safer.

Then there are horns. Horns are also microwave

emitters. Horns have a leg-thick, hollow tube called a

wave-guide at the bottom. The waveguide supplies the

microwave radiation through a hollow metallic pipe, and

the horn reflects this blast of microwave radiation off an

interior reflector, into a narrow beam of the proper

"phase," "aperture," and "directivity." Horn antennas are

narrow at the bottom and spread out at the top, like

acoustic horns. Some are conical, others rectangular.

They tend to be mounted vertically inside the tower

structure. The "noise" of the horn comes out the side of

the horn, not its end, however.

One may see a number of white poles, mounted

vertically, spaced parallel and rather far apart, attached

to the tower but well away from it. On big towers, these

poles might be half-way up; on shorter towers, they're at

the top. Sometimes the vertical poles are mounted on the

rim of a square or triangular platform, with catwalks for

easy access by tower hands. These are antennas for land

mobile radio services: paging, cellular phones, cab

dispatch, and express mail services.

The tops of towers may well be thick, pipelike,

featureless cylinders. These are generally TV broadcast

antennas encased in a long cylindrical radome, and topped

off with an aircraft beacon.

Very odd things grow from the sides of towers. One

sometimes sees a tall vertically mounted rack of metal

curlicues that look like a stack of omega signs. These

are tubular ring antennas with one knobby stub pointing

upward, one stub downward, in an array of up to sixteen.

These are FM radio transmitters.

Another array of flat metal rings is linked lengthwise

by two long parallel rods. These are VHF television

broadcast antennas.

Another species of FM antenna is particularly odd.

These witchy-looking arrays stand well out from the side

of the tower, on a rod with two large, V-shaped pairs of

arms. One V is out at the end of the rod, canted

backward, and the other is near the butt of the rod,

canted forward. The two V's are twisted at angles to one

another, so that from the ground the ends of the V's

appear to overlap slightly, forming a broken square. The

arms are of hollow brass tubing, and they come in long

sets down the side of the tower. The whole array

resembles a line of children's jacks that have all been

violently stepped on.

The four arms of each antenna are quarter-wavelength

arms, two driven and two parasitic, so that their FM

radiation is in 90-degree quadrature with equal amplitudes

and a high aperture efficiency. Of course, that's easy

for *you* to say...

In years to come, the ecology of towers will probably

change greatly. This is due to the weird phenomenon known

as the "Great Media Exchange" or the "Negroponte Flip,"

after MIT media theorist Nicholas Negroponte. Broadcast

services such as television are going into wired

distribution by cable television, where a single

"broadcast" can reach 60 percent of the American

population and even reach far overseas. With a

combination of cable television in cities and direct

satellite broadcast rurally, what real need remains for

television towers? In the meantime, however, services

formerly transferred exclusively by wire, such as

telephone and fax, are going into wireless, cellular,

portable, applications, supported by an infrastructure of

small neighborhood towers and rather modestly-sized

antennas.

Antennas have a glowing future. The spectrum can only

become more crowded, and the design of antennas can only

become more sophisticated. It may well be, though, that

another couple of decades will reduce the great steel

spires of the skyline to relics. We have seen them every

day of our lives, grown up with them as constant looming

presences. But despite their steel and their size, their

role in society may prove no more permanent than that of

windmills or lighthouses. If we do lose them to the

impetus of progress, our grandchildren will regard these

great towers with a mixture of romance and incredulity, as

the largest and most garish technological anomalies that

the twentieth century ever produced.

"The New Cryptography"

Writing is a medium of communication and

understanding, but there are times and places when one

wants an entirely different function from writing:

concealment and deliberate bafflement.

Cryptography, the science of secret writing, is almost

as old as writing itself. The hieroglyphics of ancient

Egypt were deliberately arcane: both writing and a cypher.

Literacy in ancient Egypt was hedged about with daunting

difficulty, so as to assure the elite powers of priest and

scribe.

Ancient Assyria also used cryptography, including the

unique and curious custom of "funerary cryptography."

Assyrian tombs sometimes featured odd sets of

cryptographic cuneiform symbols. The Assyrian passerby,

puzzling out the import of the text, would mutter the

syllables aloud, and find himself accidentally uttering a

blessing for the dead. Funerary cryptography was a way

to steal a prayer from passing strangers.

Julius Caesar lent his name to the famous "Caesar

cypher," which he used to secure Roman military and

political communications.

Modern cryptographic science is deeply entangled with

the science of computing. In 1949, Claude Shannon, the

pioneer of information theory, gave cryptography its

theoretical foundation by establishing the "entropy" of a

message and a formal measurement for the "amount of

information" encoded in any stream of digital bits.

Shannon's theories brought new power and sophistication to

the codebreaker's historic efforts. After Shannon,

digital machinery could pore tirelessly and repeatedly

over the stream of encrypted gibberish, looking for

repetitions, structures, coincidences, any slight

variation from the random that could serve as a weak point

for attack.

Computer pioneer Alan Turing, mathematician and

proponent of the famous "Turing Test" for artificial

intelligence, was a British cryptographer in the 1940s.

In World War II, Turing and his colleagues in espionage

used electronic machinery to defeat the elaborate

mechanical wheels and gearing of the German Enigma code-

machine. Britain's secret triumph over Nazi

communication security had a very great deal to do with

the eventual military triumph of the Allies. Britain's

code-breaking triumph further assured that cryptography

would remain a state secret and one of the most jealously

guarded of all sciences.

After World War II, cryptography became, and has

remained, one of the crown jewels of the American national

security establishment. In the United States, the science

of cryptography became the high-tech demesne of the

National Security Agency (NSA), an extremely secretive

bureaucracy that President Truman founded by executive

order in 1952, one of the chilliest years of the Cold War.

Very little can be said with surety about the NSA.

The very existence of the organization was not publicly

confirmed until 1962. The first appearance of an NSA

director before Congress was in 1975. The NSA is said to

be based in Fort Meade, Maryland. It is said to have a

budget much larger than that of the CIA, but this is

impossible to determine since the budget of the NSA has

never been a matter of public record. The NSA is said to

the the largest single employer of mathematicians in the

world. The NSA is estimated to have about 40,000

employees. The acronym NSA is aptly said to stand for

"Never Say Anything."

The NSA almost never says anything publicly. However,

the NSA's primary role in the shadow-world of electronic

espionage is to protect the communications of the US

government, and crack those of the US government's real,

imagined, or potential adversaries. Since this list of

possible adversaries includes practically everyone, the

NSA is determined to defeat every conceivable

cryptographic technique. In pursuit of their institutional

goal, the NSA labors (in utter secrecy) to crack codes and

cyphers and invent its own less breakable ones.

The NSA also tries hard to retard civilian progress in

the science of cryptography outside its own walls. The

NSA can suppress cryptographic inventions through the

little-known but often-used Invention Secrecy Act of 1952,

which allows the Commissioner of Patents and Trademarks to

withhold patents on certain new inventions and to order

that those inventions be kept secret indefinitely, "as the

national interest requires." The NSA also seeks to

control dissemination of information about cryptography,

and to control and shape the flow and direction of

civilian scientific research in the field.

Cryptographic devices are formally defined as

"munitions" by Title 22 of the United States Code, and are

subject to the same import and export restrictions as

arms, ammunition and other instruments of warfare.

Violation of the International Traffic in Arms Regulations

(ITAR) is a criminal affair investigated and administered

by the Department of State. It is said that the

Department of State relies heavily on NSA expert advice in

determining when to investigate and/or criminally

prosecute illicit cryptography cases (though this too is

impossible to prove).

The "munitions" classification for cryptographic

devices applies not only to physical devices such as

telephone scramblers, but also to "related technical data"

such as software and mathematical encryption algorithms.

This specifically includes scientific "information" that

can be "exported" in all manner of ways, including simply

verbally discussing cryptography techniques out loud. One

does not have to go overseas and set up shop to be

regarded by the Department of State as a criminal

international arms trafficker. The security ban

specifically covers disclosing such information to any

foreign national anywhere, including within the borders of

the United States.

These ITAR restrictions have come into increasingly

harsh conflict with the modern realities of global

economics and everyday real life in the sciences and

academia. Over a third of the grad students in computer

science on American campuses are foreign nationals.

Strictly appled ITAR regulations would prevent

communication on cryptography, inside an American campus,

between faculty and students. Most scientific journals

have at least a few foreign subscribers, so an exclusively

"domestic" publication about cryptography is also

practically impossible. Even writing the data down on a

cocktail napkin could be hazardous: the world is full of

photocopiers, modems and fax machines, all of them

potentially linked to satellites and undersea fiber-optic

cables.

In the 1970s and 1980s, the NSA used its surreptitious

influence at the National Science Foundation to shape

scientific research on cryptography through restricting

grants to mathematicians. Scientists reacted mulishly, so

in 1978 the Public Cryptography Study Group was founded as

an interface between mathematical scientists in civilian

life and the cryptographic security establishment. This

Group established a series of "voluntary control"

measures, the upshot being that papers by civilian

researchers would be vetted by the NSA well before any

publication.

This was one of the oddest situations in the entire

scientific enterprise, but the situation was tolerated for

years. Most US civilian cryptographers felt, through

patriotic conviction, that it was in the best interests of

the United States if the NSA remained far ahead of the

curve in cryptographic science. After all, were some

other national government's electronic spies to become

more advanced than those of the NSA, then American

government and military transmissions would be cracked and

penetrated. World War II had proven that the

consequences of a defeat in the cryptographic arms race

could be very dire indeed for the loser.

So the "voluntary restraint" measures worked well for

over a decade. Few mathematicians were so enamored of

the doctrine of academic freedom that they were prepared

to fight the National Security Agency over their supposed

right to invent codes that could baffle the US government.

In any case, the mathematical cryptography community was a

small group without much real political clout, while the

NSA was a vast, powerful, well-financed agency

unaccountable to the American public, and reputed to

possess many deeply shadowed avenues of influence in the

corridors of power.

However, as the years rolled on, the electronic

exchange of information became a commonplace, and users of

computer data became intensely aware of their necessity

for electronic security over transmissions and data. One

answer was physical security -- protect the wiring, keep

the physical computers behind a physical lock and key.

But as personal computers spread and computer networking

grew ever more sophisticated, widespread and complex, this

bar-the-door technique became unworkable.

The volume and importance of information transferred

over the Internet was increasing by orders of magnitude.

But the Internet was a notoriously leaky channel of

information -- its packet-switching technology meant that

packets of vital information might be dumped into the

machines of unknown parties at almost any time. If the

Internet itself could not locked up and made leakproof --

and this was impossible by the nature of the system --

then the only secure solution was to encrypt the message

itself, to make that message unusable and unreadable, even

if it sometimes fell into improper hands.

Computers outside the Internet were also at risk.

Corporate computers faced the threat of computer-intrusion

hacking, from bored and reckless teenagers, or from

professional snoops and unethical business rivals both

inside and outside the company. Electronic espionage,

especially industrial espionage, was intensifying. The

French secret services were especially bold in this

regard, as American computer and aircraft executives found

to their dismay as their laptops went missing during Paris

air and trade shows. Transatlantic commercial phone calls

were routinely tapped by French government spooks seeking

commercial advantage for French companies in the computer

industry, aviation, and the arms trade. And the French

were far from alone when it came to government-supported

industrial espionage.

Protection of private civilian data from foreign

government spies required that seriously powerful

encryption techniques be placed into private hands.

Unfortunately, an ability to baffle French spies also

means an ability to baffle American spies. This was not

good news for the NSA.

By 1993, encryption had become big business. There

were one and half million copies of legal encryption

software publicly available, including widely-known and

commonly-used personal computer products such as Norton

Utilities, Lotus Notes, StuffIt, and several Microsoft

products. People all over the world, in every walk of

life, were using computer encryption as a matter of

course. They were securing hard disks from spies or

thieves, protecting certain sections of the family

computer from sticky-fingered children, or rendering

entire laptops and portables into a solid mess of

powerfully-encrypted Sanskrit, so that no stranger could

walk off with those accidental but highly personal life-

histories that are stored in almost every PowerBook.

People were no longer afraid of encryption.

Encryption was no longer secret, obscure, and arcane;

encryption was a business tool. Computer users wanted

more encryption, faster, sleeker, more advanced, and

better.

The real wild-card in the mix, however, was the new

cryptography. A new technique arose in the 1970s:

public-key cryptography. This was an element the

codemasters of World War II and the Cold War had never

foreseen.

Public-key cryptography was invented by American

civilian researchers Whitfield Diffie and Martin Hellman,

who first published their results in 1976.

Conventional classical cryptographic systems, from the

Caesar cipher to the Nazi Enigma machine defeated by Alan

Turing, require a single key. The sender of the message

uses that key to turn his plain text message into

cyphertext gibberish. He shares the key secretly with the

recipients of the message, who use that same key to turn

the cyphertext back into readable plain text.

This is a simple scheme; but if the key is lost to

unfriendly forces such as the ingenious Alan Turing, then

all is lost. The key must therefore always remain hidden,

and it must always be fiercely protected from enemy

cryptanalysts. Unfortunately, the more widely that key is

distributed, the more likely it is that some user in on

the secret will crack or fink. As an additional burden,

the key cannot be sent by the same channel as the

communications are sent, since the key itself might be

picked-up by eavesdroppers.

In the new public-key cryptography, however, there are

two keys. The first is a key for writing secret text,

the second the key for reading that text. The keys are

related to one another through a complex mathematical

dependency; they determine one another, but it is

mathematically extremely difficult to deduce one key from

the other.

The user simply gives away the first key, the "public

key," to all and sundry. The public key can even be

printed on a business card, or given away in mail or in a

public electronic message. Now anyone in the public, any

random personage who has the proper (not secret, easily

available) cryptographic software, can use that public key

to send the user a cyphertext message. However, that

message can only be read by using the second key -- the

private key, which the user always keeps safely in his own

possession.

Obviously, if the private key is lost, all is lost.

But only one person knows that private key. That private

key is generated in the user's home computer, and is never

revealed to anyone but the very person who created it.

To reply to a message, one has to use the public key

of the other party. This means that a conversation

between two people requires four keys. Before computers,

all this key-juggling would have been rather unwieldy, but

with computers, the chips and software do all the

necessary drudgework and number-crunching.

The public/private dual keys have an interesting

alternate application. Instead of the public key, one can

use one's private key to encrypt a message. That message

can then be read by anyone with the public key, i.e,.

pretty much everybody, so it is no longer a "secret"

message at all. However, that message, even though it is

no longer secret, now has a very valuable property: it is

authentic. Only the individual holder of the private key

could have sent that message.

This authentication power is a crucial aspect of the

new cryptography, and may prove to be more socially

important than secrecy. Authenticity means that

electronic promises can be made, electronic proofs can be

established, electronic contracts can be signed,

electronic documents can be made tamperproof. Electronic

impostors and fraudsters can be foiled and defeated -- and

it is possible for someone you have never seen, and will

never see, to prove his bona fides through entirely

electronic means.

That means that economic relations can become

electronic. Theoretically, it means that digital cash is

possible -- that electronic mail, e-mail, can be joined by

a strange and powerful new cousin, electronic cash, e-

money.

Money that is made out of text -- encrypted text. At

first consideration such money doesn't seem possible,

since it is so far outside our normal experience. But

look at this:

ASCII-picture of US dollar

This parody US banknote made of mere letters and

numbers is being circulated in e-mail as an in-joke in

network circles. But electronic money, once established,

would be no more a joke than any other kind of money.

Imagine that you could store a text in your computer and

send it to a recipient; and that once gone, it would be

gone from your computer forever, and registered infallibly

in his. With the proper use of the new encryption and

authentication, this is actually possible. Odder yet, it

is possible to make the note itself an authentic, usable,

fungible, transferrable note of genuine economic value,

without the identity of its temporary owner ever being

made known to anyone. This would be electronic cash --

like normal cash, anonymous -- but unlike normal cash,

lightning-fast and global in reach.

There is already a great deal of electronic funds

transfer occurring in the modern world, everything from

gigantic currency-exchange clearinghouses to the

individual's VISA and MASTERCARD bills. However, charge-

card funds are not so much "money" per se as a purchase

via proof of personal identity. Merchants are willing to

take VISA and MASTERCARD payments because they know that

they can physically find the owner in short order and, if

necessary, force him to pay up in a more conventional

fashion. The VISA and MASTERCARD user is considered a

good risk because his identity and credit history are

known.

VISA and MASTERCARD also have the power to accumulate

potentially damaging information about the commercial

habits of individuals, for instance, the video stores one

patronizes, the bookstores one frequents, the restaurants

one dines in, or one's travel habits and one's choice of

company.

Digital cash could be very different. With proper

protection from the new cryptography, even the world's

most powerful governments would be unable to find the

owner and user of digital cash. That cash would secured

by a "bank" -- (it needn't be a conventional, legally

established bank) -- through the use of an encrypted

digital signature from the bank, a signature that neither

the payer nor the payee could break.

The bank could register the transaction. The bank

would know that the payer had spent the e-money, and the

bank could prove that the money had been spent once and

only once. But the bank would not know that the payee had

gained the money spent by the payer. The bank could track

the electronic funds themselves, but not their location or

their ownership. The bank would guarantee the worth of

the digital cash, but the bank would have no way to tie

the transactions together.

The potential therefore exists for a new form of

network economics made of nothing but ones and zeroes,

placed beyond anyone's controls by the very laws of

mathematics. Whether this will actually happen is

anyone's guess. It seems likely that if it did happen, it

would prove extremely difficult to stop.

Public-key cryptography uses prime numbers. It is a

swift and simple matter to multiply prime numbers together

and obtain a result, but it is an exceedingly difficult

matter to take a large number and determine the prime

numbers used to produce it. The RSA algorithm, the

commonest and best-tested method in public-key

cryptography, uses 256-bit and 258-bit prime numbers.

These two large prime numbers ("p" and "q") are used to

produce very large numbers ("d" and "e") so that (de-1) is

divisible by (p-1) times (q-1). These numbers are easy to

multiply together, yielding the public key, but extremely

difficult to pull apart mathematically to yield the

private key.

To date, there has been no way to mathematically prove

that it is inherently difficult to crack this prime-number

cipher. It might be very easy to do if one knew the

proper advanced mathematical technique for it, and the

clumsy brute-power techniques for prime-number

factorization have been improving in past years. However,

mathematicians have been working steadily on prime number

factorization problems for many centuries, with few

dramatic advances. An advance that could shatter the RSA

algorithm would mean an explosive breakthrough across a

broad front of mathematical science. This seems

intuitively unlikely, so prime-number public keys seem

safe and secure for the time being -- as safe and secure

as any other form of cryptography short of "the one-time

pad." (The one-time pad is a truly unbreakable cypher.

Unfortunately it requires a key that is every bit as long

as the message, and that key can only be used once. The

one-time pad is solid as Gibraltar, but it is not much

practical use.)

Prime-number cryptography has another advantage. The

difficulty of factorizing numbers becomes drastically

worse as the prime numbers become larger. A 56-bit key

is, perhaps, not entirely outside the realm of possibility

for a nationally supported decryption agency with large

banks of dedicated supercomputers and plenty of time on

their hands. But a 2,048 bit key would require every

computer on the planet to number-crunch for hundreds of

centuries.

Decrypting a public-keyed message is not so much a

case of physical impossibility, as a matter of economics.

Each key requires a huge computational effort to break it,

and there are already thousands of such keys used by

thousands of people. As a further blow against the

decryptor, the users can generate new keys easily, and

change them at will. This poses dire problems for the

professional electronic spy.

The best-known public-key encryption technique, the

RSA algorithm, was named after its inventors, Ronald L.

Rivest, Adi Shamir and Leon Adleman. The RSA technique

was invented in the United States in the late 1980s

(although, as if to spite the international trade in arms

regulations, Shamir himself is an Israeli). The RSA

algorithm is patented in the United States by the

inventors, and the rights to implement it on American

computers are theoretically patented by an American

company known as Public Key Partners. (Due to a patent

technicality, the RSA algorithm was not successfully

patented overseas.)

In 1991 an amateur encryption enthusiast named Phil

Zimmerman wrote a software program called "Pretty Good

Privacy" that used the RSA algorithm without permission.

Zimmerman gave the program away on the Internet network

via modem from his home in Colorado, because of his

private conviction that the public had a legitimate need

for powerful encryption programs at no cost (and,

incidentally, no profit to the inventors of RSA). Since

Zimmerman's action, "Pretty Good Privacy" or "PGP" has

come into common use for encrypting electronic mail and

data, and has won an avid international following. The

original PGP program has been extensively improved by

other software writers overseas, out of the reach of

American patents or the influence of the NSA, and the PGP

program is now widely available in almost every country on

the planet -- or at least, in all those countries where

floppy disks are common household objects.

Zimmerman, however, failed to register as an arms

dealer when he wrote the PGP software in his home and made

it publicly available. At this writing, Zimmerman is

under federal investigation by the Office of Defense Trade

Controls at the State Department, and is facing a

possible criminal indictment as an arms smuggler. This

despite the fact that Zimmerman was not, in fact, selling

anything, but rather giving software away for free. Nor

did he voluntarily "export" anything -- rather, people

reached in from overseas via Internet links and retrieved

Zimmerman's program from the United States under their own

power and through their own initiative.

Even more oddly, Zimmerman's program does not use the

RSA algorithm exclusively, but also depends on the

perfectly legal DES or Data Encryption Standard. The Data

Encryption Standard, which uses a 56-bit classical key,

is an official federal government cryptographic technique,

created by IBM with the expert help of the NSA. It has

long been surmised, though not proven, that the NSA can

crack DES at will with their legendary banks of Cray

supercomputers. Recently a Canadian mathematician,

Michael Wiener of Bell-Northern Research, published plans

for a DES decryption machine that can purportedly crack

56-bit DES in a matter of hours, through brute force

methods. It seems that the US Government's official 56-

bit key -- insisted upon, reportedly, by the NSA -- is now

too small for serious security uses.

The NSA, and the American law enforcement community

generally, are unhappy with the prospect of privately

owned and powerfully secure encryption. They acknowledge

the need for secure communications, but they insist on the

need for police oversight, police wiretapping, and on the

overwhelming importance of national security interests and

governmental supremacy in the making and breaking of

cyphers.

This motive recently led the Clinton Administration to

propose the "Clipper Chip" or "Skipjack," a government-

approved encryption device to be placed in telephones.

Sets of keys for the Clipper Chip would be placed in

escrow with two different government agencies, and when

the FBI felt the need to listen in on an encrypted

telephone conversation, the FBI would get a warrant from a

judge and the keys would be handed over.

Enthusiasts for private encryption have pointed out a

number of difficulties with the Clipper Chip proposal.

First of all, it is extremely unlikely that criminals,

foreign spies, or terrorists would be foolish enough to

use an encryption technique designed by the NSA and

approved by the FBI. Second, the main marketing use for

encryption is not domestic American encryption, but

international encryption. Serious business users of

serious encryption are far more alarmed by state-supported

industrial espionage overseas, than they are about the

safety of phone calls made inside the United States. They

want encryption for communications made overseas to people

overseas -- but few foreign business people would buy an

encryption technology knowing that the US Government held

the exclusive keys.

It is therefore likely that the Clipper Chip could

never be successfully exported by American manufacturers

of telephone and computer equipment, and therefore it

could not be used internationally, which is the primary

market for encryption. Machines with a Clipper Chip

installed would become commercial white elephants, with no

one willing to use them but American cops, American spies,

and Americans with nothing to hide.

A third objection is that the Skipjack algorithm has

been classified "Secret" by the NSA and is not available

for open public testing. Skeptics are very unwilling to

settle for a bland assurance from the NSA that the chip

and its software are unbreakable except with the official

keys.

The resultant controversy was described by Business

Week as "Spy Vs Computer Nerd." A subterranean power-

struggle has broken out over the mastery of cryptographic

science, and over basic ownership of the electronic bit-

stream.

Much is riding on the outcome.

Will powerful, full-fledged, state-of-the-art

encryption belong to individuals, including such unsavory

individuals as drug traffickers, child pornographers,

black-market criminal banks, tax evaders, software

pirates, and the possible future successors of the Nazis?

Or will the NSA and its allies in the cryptographic

status-quo somehow succeed in stopping the march of

scientific progress in cryptography, and in cramming the

commercial crypto-genie back into the bottle? If so, what

price will be paid by society, and what damage wreaked on

our traditions of free scientific and technical inquiry?

One thing seems certain: cryptography, this most

obscure and smothered of mathematical sciences, is out in

the open as never before in its long history.

Impassioned, radicalized cryptographic enthusiasts, often

known as "cypherpunks," are suing the NSA and making it

their business to spread knowledge of cryptographic

techniques as widely as possible, "through whatever means

necessary." Small in number, they nevertheless have

daring, ingenuity, and money, and they know very well how

to create a public stink. In the meantime, their more

conventional suit-and-tie allies in the Software

Publishers Association grumble openly that the Clipper

Chip is a poorly-conceived fiasco, that cryptographic

software is peddled openly all over the planet, and that

"the US Government is succeeding only in crippling an

American industry's exporting ability."

The NSA confronted the worst that America's

adversaries had to offer during the Cold War, and the NSA

prevailed. Today, however, the secret masters of

cryptography find themselves confronting what are perhaps

the two most powerful forces in American society: the

computer revolution, and the profit motive. Deeply hidden

from the American public through forty years of Cold War

terror, the NSA itself is for the first time, exposed to

open question and harrowing reassessment.

Will the NSA quietly give up the struggle, and expire

as secretly and silently as it lived its forty-year Cold

War existence? Or will this most phantomlike of federal

agencies decide to fight for its survival and its

scientific pre-eminence?

And if this odd and always-secret agency does choose

to fight the new cryptography, then -- how?

"The Dead Collider"

It certainly seemed like a grand idea at the time, the

time being 1982, one of the break-the-bank years of the

early Reagan Administration.

The Europeans at CERN, possessors of the world's

largest particle accelerator, were planning to pave their

massive Swiss tunnel with new, superconducting magnets.

This would kick the European atom-smasher, already

powerful, up to a massive 10 trillion electron volts.

In raw power, this would boost the Europeans

decisively past their American rivals. America's most

potent accelerator in 1982, Fermilab in Illinois, could

manage a meager 2 TeV. And Fermilab's Tevatron, though

upgraded several times, was an aging installation.

A more sophisticated machine, ISABELLE at Brookhaven

National Laboratory in New York, had been planned in 1979

as Fermilab's successor at the forefront of American

particle physics. But by 1982, it was clear that

ISABELLE's ultra-sophisticated superconducting magnets had

severe design troubles. The state-of-the-art bungling at

Brookhaven was becoming an open embarrassment to the

American particle-physics community. And even if the

young ISABELLE facility overcame those problems and got

their magnets to run, ISABELLE was intended to sacrifice

raw power for sophistication; at best, ISABELLE would

yield a feeble .8 TeV.

In August 1982, Leon Lederman, then director of

Fermilab, made a bold and visionary proposal. In a

conference talk to high-energy physicists gathered in

Colorado, Lederman proposed cancelling both ISABELLE and

the latest Fermilab upgrade, in pursuit of a gigantic

American particle accelerator that would utterly dwarf the

best the Europeans had to offer, now or in the foreseeable

future. He called it "The Machine in the Desert."

The "Desertron" (as Lederman first called it) would be

the largest single scientific instrument in the world,

employing a staff of more than two thousand people, plus

students, teachers and various properly awestruck visiting

scholars from overseas. It would be 20 times more

powerful than Fermilab, and full sixty times more powerful

than CERN circa 1982. The accelerator's 54 miles of deep

tunnels, lined with hard- vacuum beamguides and helium-

refrigerated giant magnets, would be fully the size of the

Washington Beltway.

The cost: perhaps 3 billion dollars. It was thought

that the cash- flush Japanese, who had been very envious

of CERN for some time, would be willing to help the

Americans in exchange for favored status at the complex.

The goal of the Desertron, or at least its target of

choice, would be the Higgs scalar boson, a hypothetical

subatomic entity theoretically responsible for the fact

that other elementary particles have mass. The Higgs

played a prominent part at the speculative edges of

quantum theory's so-called "Standard Model," but its true

nature and real properties were very much in doubt.

The Higgs boson would be a glittering prize indeed,

though not so glittering as the gigantic lab itself.

After a year of intense debate within the American high-

energy-physics community, Lederman's argument won out.

His reasoning was firmly in the tradition of 20th-

century particle physics. There seemed little question

that massive power and scale of the Desertron was the

necessary next step for real progress in the field.

At the beginning of the 20th century, Ernest

Rutherford (who coined the memorable catch-phrase, "All

science is either physics or stamp-collecting") discovered

the nucleus of the atom with a mere five million electron

volts. Rutherford's lab equipment not much more

sophisticated than string and sealing-wax. To get

directly at neutrons and protons, however, took much more

energy -- a billion electron volts and a cyclotron. To

get quark effects, some decades later, required ten

billion volts and a synchrotron. To make quarks really

stand up and dance in their full quantum oddity, required

a hundred billion electron volts and a machine that was

miles across. And to get at the Higgs boson would need

at least ten trillion eV, and given that the fantastically

powerful collision would be a very messy affair, a full

forty trillion -- two particle beams of twenty TeV each,

colliding head-on -- was a much safer bet.

Throughout the century, then, every major new advance

in particle studies had required massive new infusions of

power. A machine for the 1990s, the end result of

decades of development, would require truly titanic

amounts of juice. The physics community had hesitated at

this step, and had settled for years at niggling around in

the low trillions of electron volts. But the field of

sub-atomic studies was looking increasingly mined-out, and

the quantum Standard Model had not had a good paradigm-

shattering kick in the pants in some time. From the

perspective of the particle physicist, the Desertron,

despite its necessarily colossal scale, made perfect

scientific sense.

The Department of Energy, the bureaucratic descendant

of the Atomic Energy Commission and the traditional

federal patron of high-energy physics, had more or less

recovered from its last major money-wasting debacle, the

Carter Administration's synthetic fuels program. Under

new leadership, the DoE was sympathetic to an ambitious

project with some workable and sellable rationale.

Lederman's tentative scheme was developed, over three

years, in great detail, by an expert central design group

of federally-sponsored physicists and engineers from

Lawrence Berkeley labs, Brookhaven and Fermilab. The

"Desertron" was officially renamed the "Superconducting

Super Collider." In 1986 the program proposal was carried

to Ronald Reagan, then in his second term. While Reagan's

cabinet seemed equally split on the merits of the SSC

versus a much more modest research program, the Gipper

decided the issue with one of his favorite football

metaphors: "Throw deep."

Reagan's SSC was a deep throw indeed. The collider

ring of Fermilab in Illinois was visible from space, and

the grounds of Fermilab were big enough to boast their

own herd of captive buffalo. But the ring of the mighty

Super Collider made Fermilab's circumference look like a

nickel on a dinner plate. One small section of the Super

Collider, the High Energy Booster, was the size of

Fermilab all by itself, but this Booster was only a

humble injection device for the Super Collider.

The real action was to be in the fifty-four-mile, 14-

ft-diameter Super Collider ring.

As if this titanic underground circus were not enough,

the SSC also boasted two underground halls each over 300

feet long, to be stuffed with ultrasophisticated particle

detectors so huge as to make their hard-helmeted minders

resemble toy dolls. Along with the fifty-four miles of

Collider were sixteen more miles of injection devices:

the Linear Accelerator, the modest Low Energy Booster, the

large Medium Energy Booster, the monster High Energy

Booster, the Boosters acting like a set of gears to drive

particles into ever-more frenzied states of relativistic

overdrive, before their release into the ferocious grip of

the main Super Collider ring.

Along the curves and arcs of these wheels-within-

wheels, and along the Super Collider ring itself, were

more than forty vertical access shafts, some of them two

hundred feet deep. Up on the surface, twelve separate

refrigeration plants would pipe tons of ultra-frigid

liquid helium to more than ten thousand superconducting

magnets, buried deep within the earth. All by itself, the

SSC would more than double the amount of helium

refrigeration taking place in the entire planet.

The site would have miles of new-paved roads, vast

cooling ponds of fresh water, brand-new electrical

utilities. Massive new office complexes were to be built

for support and research, including two separate East and

West campuses at opposite ends of the Collider, and two

offsite research labs. With thousands of computers:

personal computers, CAD workstations, network servers,

routers, massively parallel supercomputing simulators.

Office and laboratory networking including Internet and

videoconferencing. Assembly buildings, tank farms,

archives, libraries, security offices, cafeterias. The

works.

There were, of course, dissenters from the dream.

Some physicists feared that the project, though workable

and probably quite necessary for any real breakthrough in

their field, was simply too much to ask. Enemies from

outside the field likened the scheme to Reagan's Star Wars

-- an utter scientific farce -- and to the Space Station,

a political pork-barrel effort with scarcely a shred of

real use in research -- and to the hapless Space Shuttle,

an overdesigned gobboon.

Within the field of high-energy-physics, though, the

logic was too compelling and the traditional arc of

development too strong. A few physicists -- Freeman Dyson

among them -- quietly suggested that it might be time for

a radically new tack; time to abandon the tried-and-true

collider technology entirely, to try daringly novel,

small-scale particle-acceleration schemes such as free-

electron lasers, gyroklystrons, or wake- field

accelerators. But that was not Big Thinking; and

particle physics was the very exemplar of Big Science.

In the 1920 and 1930s, particle physicist Ernest

Lawrence had practically invented "Big Science" with the

Berkeley cyclotrons, each of them larger, more expensive,

demanding greater resources and entire teams of

scientists. Particle physics, in pursuit of ever-more-

elusive particles, by its nature built huge, centralized

facilities of ever greater complexity and ever greater

expense for ever-larger staffs of researchers. There

just wasn't any other way to do particle physics, but the

big way.

And then there was the competitive angle, the race for

international prestige: high-energy physics as the

arcane, scholarly equivalent of the nuclear arms race.

The nuclear arms race itself was, of course, a direct

result of progress in 20th-century high-energy physics.

For Cold Warriors, nuclear science, with its firm linkage

to military power, was the Big Science par excellence.

Leon Lederman and his colleague Sheldon Glashow played

the patriotic card very strongly in their influential

article of March 1985, "The SSC: A Machine for the

Nineties." There they wrote: "Of course, as scientists,

we must rejoice in the brilliant achievements of our

colleagues overseas. Our concern is that if we forgo the

opportunity that SSC offers for the 1990s, the loss will

not only be to our science but also to the broader issue

of national pride and technological self-confidence. When

we were children, America did most things best. So it

should again."

Lederman and Glashow also argued for the SSC on the

grounds of potential spinoffs for American industry:

energy storage, power transmission, new tunneling

techniques, industrial demand-pull in superconductivity.

In meeting "all but insuperable technical obstacles," they

declared, American industries would learn better to

compete. (There was no mention of what might happen to

American "national pride and technological self-

confidence" if American industries simply failed to meet

those "insuperable obstacles" -- as had already happened

in ISABELLE.)

Glashow and Lederman also declared, with perhaps

pardonable professional pride, that it was simply a good

idea for America to create and employ large armies of

particle physicists, pretty much for their own sake.

"(P)article physics yields highly trained scientists

accustomed to solving the unsolvable. They often go on to

play vital roles in the rest of the world.... Many of us

have become important contributors in the world of energy

resources, neurophysiology, arms control and disarmament,

high finance, defense technology and molecular biology....

High energy physics continues to attract and recruit into

science its share of the best and brightest. If we were

deprived of all those who began their careers with the

lure and the dream of participating in this intellectual

adventure, the nation would be considerably worse off than

it is. Without the SSC, this is exactly what would come

to pass."

Funding a gigantic physics lab may seem a peculiarly

roundabout way to create, say, molecular biologists,

especially when America's actual molecular biologists, no

slouches at "solving the unsolvable" themselves, were

getting none of the funding for the Super Collider.

When it came to creating experts in "high finance,"

however, the SSC was on much firmer ground. Financiers

worked overtime as the SSC's cost estimates rose again and

again, in leaps of billions. The Japanese were quite

interested in basic research in superconductive

technology; but when they learned they were expected to

pay a great deal, but enjoy little of the actual technical

development in superconductivity, they naturally balked.

So did the Taiwanese, when an increasingly desperate SSC

finally got around to asking them to help. The

Europeans, recognizing a direct attempt to trump their

treasured CERN collider, were superconductively chilly

about the idea of investing in any Yankee dream- machine.

Estimated cost of the project to the American taxpayer --

or rather, the American deficit borrower -- quickly jumped

from 3.9 billion dollars to 4.9 billion, then 6.6

billion, then 8.25 billion, then 10 billion. Then,

finally and fatally, to twelve.

Time and again the physicists went to the

Congressional crap table, shot the dice for higher stakes,

and somehow survived. Scientists outside the high-energy-

physics community were livid with envy, but the powerful

charisma of physics -- that very well-advanced field that

had given America the atomic bomb and a raft of Nobels --

held firm against the jealous, increasingly bitter gaggle

of "little science" advocates.

At the start of the project, the Congress was highly

enthusiastic. The lucky winner of the SSC had a great

deal to gain: a nucleus of high-tech development,

scientific prestige, and billions in federally-subsidized

infrastructure investment. The Congressperson carrying

the SSC home to the district would have a prize beyond

mere water-project pork; that lucky politician would have

trapped a mastodon.

At length the lucky winner of the elaborate site-

selection process was announced: Waxahachie, Texas.

Texas Congresspeople were, of course, ecstatic; but other

competitors wondered what on earth Waxahachie had to offer

that they couldn't.

Waxahachie's main appeal was simple: lots of Texas-

sized room for a Texas-sized machine. The Super Collider

would, in fact, entirely encircle the historic town of

Waxahachie, some 18,000 easy-going folks in a rural county

previously best known for desultory cotton-farming. The

word "Waxahachie" originally meant "buffalo creek."

Waxahachie was well-watered, wooded, farming country built

on a bedrock of soft, chalky, easily-excavated limestone.

Lederman, author of the Desertron proposal, rudely

referred to Waxahachie as being "in Texas, in the desert"

in his SSC promotional pop- science book THE GOD PARTICLE.

There was no desert anywhere near Waxahachie, and worse

yet, Lederman had serious problems correctly pronouncing

the town's name.

The town of Waxahachie, a minor railroad boomtown in

the 1870s and 1880s, had changed little during the

twentieth century. In later years, Waxahachie had made a

virtue of its fossilization. Downtown Waxahachie had a

striking Victorian granite county courthouse and a brick-

and- gingerbread historical district of downtown shops,

mostly frequented by antique-hunting yuppies on day-

trips from the Dallas-Fort Worth Metroplex, twenty miles

to the north. There was a certain amount of suburban

sprawl on the north edge of town, at the edge of commuting

range to south Dallas, but it hadn't affected the pace of

local life much. Quiet, almost sepulchral Waxahachie was

the most favored place in Texas for period moviemaking.

Its lovely oak-shadowed graveyard was one of the most-

photographed cemeteries in the entire USA.

This, then, was to become the new capital of the

high-energy physics community, the home of a global

scientific community better known for Mozart and chablis

than catfish and C&W. It seemed unbelievable. And it was

unbelievable. Scientifically, Waxahachie made sense.

Politically, Waxahachie could be sold. Culturally,

Waxahachie made no sense whatsoever. A gesture by the

federal government and a giant machine could not, in fact,

transform good ol' Waxahachie into Berkeley or Chicago or

Long Island. A mass migration of physicists might have

worked for Los Alamos when hundreds of A-Bomb scientists

had been smuggled there in top secrecy at the height of

World War II, but there was no atomic war on at the

moment. A persistent sense of culture shock and

unreality haunted the SSC project from the beginning.

In his 1993 popular-science book THE GOD PARTICLE,

Lederman made many glowing comparisons for the SSC: the

cathedrals of Europe, the Pyramids, Stonehenge. But those

things could all be seen. They all made instant sense

even to illiterates. The SSC, unlike the Pyramids, was

almost entirely invisible -- a fifty-mile subterranean

wormhole stuffed with deep-frozen magnets.

A trip out to the SSC revealed construction cranes,

vast junkyards of wooden crating and metal piping, with a

few drab, rectangular, hopelessly unromantic assembly

buildings, buildings with all the architectural vibrancy

of slab-sided machine-shops (which is what they were).

Here and there were giant weedy talus-heaps of limestone

drill-cuttings from the subterranean "TBM," or Tunnel

Boring Machine. The Boring Machine was a state-of-the-art

Boring Machine, but its workings were invisible to all but

the hard-hats, and the machine itself was, well, boring.

Here and there along the SSC's fifty-four mile

circumference, inexplicable white vents rose from the

middle of muddy cottonfields. These were the SSC's

ventilation and access shafts, all of them neatly

padlocked in case some mischievous soul should attempt to

see what all the fuss was about. Nothing at the SSC was

anything like the heart-lifting spires of Notre Dame, or

even the neat-o high-tech blast of an overpriced and

rickety Space Shuttle. The place didn't look big or

mystical or uplifting; it just looked dirty and flat and

rather woebegone.

As a popular attraction the SSC was a bust; and time

was not on the side of its planners and builders. As the

Cold War waned, the basic prestige of nuclear physics was

also wearing rather thin. Hard times had hit America,

and hard times had come for American science.

Lederman himself, onetime chairman of the board of the

American Association for the Advancement of Science, was

painfully aware of the sense of malaise and decline. In

1990 and 1991, Lederman, as chairman of AAAS, polled his

colleagues in universities across America about the basic

state of Science in America. He heard, and published, a

great outpouring of discontent. There was a litany of

complaint from American scholars. Pernickety government

oversight. Endless paperwork for grants, consuming up to

thirty percent of a scientist's valuable research time. A

general aging of the academic populace, with graying

American scientists more inclined to look back to vanished

glories than to anticipate new breakthroughs.

Meanspirited insistence by both government and industry

that basic research show immediate and tangible economic

benefits. A loss of zest and interest in the future,

replaced by a smallminded struggle to keep making daily

ends meet.

It was getting hard to make a living out there. The

competition for money and advancement inside science was

getting fierce, downright ungentlemanly. Big wild dreams

that led to big wild breakthroughs were being nipped in

the bud by a general societal malaise and a failure of

imagination. The federal research effort was still vast

in scope, and had been growing steadily despite the

steadily growing federal deficits. But thanks to decades

of generous higher education and the alluring prestige of

a life in research, there were now far more mouths to feed

in the world of Science. Vastly increased armies of grad

students and postdocs found themselves waiting forever for

tenure. They were forced to play careerist games over

shrinking slices of the grantsmanship pie, rather than

leaving money problems to the beancounters and getting

mano-a-mano with the Big Questions.

"The 1950s and 1960s were great years for science in

America," Lederman wrote nostalgically. "Compared to the

much tougher 1990s, anyone with a good idea and a lot of

determination, it seemed, could get his idea funded.

Perhaps this is as good a criterion for healthy science as

any." By this criterion, American science in the 90s was

critically ill. The SSC seemed to offer a decisive way to

break out of the cycle of decline, to return to those good

old days. The Superconducting Super Collider would make

Big Science really "super" again, not just once but twice.

The death of the project was slow, and agonizing, and

painful. Again and again particle physicists went to

Congress to put their hard-won prestige on the line, and

their supporters used every tactic in the book. As

SCIENCE magazine put in a grim postmortem editorial: "The

typical hide-and-seek game of 'it's not the science, it's

the jobs' on Monday, Wednesday, and Friday and 'it's not

about jobs, it is very good science' on Tuesday, Thursday

and Saturday wears thin after a while."

The House killed the Collider in June 1992; the Senate

resurrected it. The House killed it again in June 1993,

the Senate once again puffed the breath of life into the

corpse, but Reagan and Bush were out of power now.

Reagan had supported SSC because he was, in his own

strange way, a visionary; Bush, though usually more

prudent, took care to protect his Texan political base.

Bush did in fact win Texas in the presidential election of

1992, but winning Texas was not enough. The party was

over. In October 1993 the Super Collider was killed yet

again. And this time it stayed dead.

In January 1994 I went to Waxahachie to see the dead

Collider.

To say that morale is low at the SSC Labs does not

begin to capture the sentiment there. Morale is

subterranean. There are still almost two thousand people

employed at the dead project; not because they have

anything much to do there, but because there is still a

tad of funding left for them to consume -- a meager six

hundred million or so. And they also stay because,

despite their alleged facility at transforming themselves

into neurophysiologists, arms control advocates, et al.,

there is simply not a whole lot of market demand anywhere

for particle physicists, at the moment.

The Dallas offices of the SSC Lab are a giant maze of

cubicles, every one of them without exception sporting a

networked color Macintosh. Employees have pinned up

xeroxed office art indicative of their mood. One was a

chart called:

"THE SIX PHASES OF A PROJECT: I. Enthusiasm.

II. Disillusionment. III. Panic. IV. Search for the

Guilty. V. Punishment of the Innocent. VI. Praise &

Honor for the Nonparticipants."

According to the chart, the SSC is now at Phase Five,

and headed for Six.

SSC staffers have a lot of rather dark jokes now.

"The Sour Grapes Alert" reads "This is a special

announcement for Supercollider employees only!! Your job

is a test. It is only a test!! Had your job been an

actual job, you would have received raises, promotions,

and other signs of appreciation!! We now return you to

your miserable state of existence."

Outside the office building, one of the lab's

monstrous brown trash dumpsters has been renamed

"Superconductor." The giant steel trash-paper compactor

does look oddly like one of the SSC's fifty-foot-long

superconducting magnets; but the point, of course, is that

trash and the magnet are now roughly equivalent in worth.

The SSC project to date has cost about two billion

dollars. Some $440,885,853 of that sum was spent by the

State of Texas, and the Governor of the State of Texas,

the volatile Ann Richards, is not at all happy about it.

The Governor's Advisory Committee on the

Superconducting Super Collider held its first meeting at

the SSC Laboratory in Dallas, on January 14, 1994. The

basic assignment of this blue-ribbon panel of Texan

scholars and politicians is to figure out how to recoup

something for Texas from this massive failed investment.

Naturally I made it my business to attend, and sat in

on a day's worth of presentations by such worthies as Bob

White, President of the National Academy of Engineering;

John Peoples, the SSC's current director; Roy Schwitters,

the SSC's original Director, who resigned in anguish after

the cancellation; the current, and former, Chancellors of

the University of Texas System; the Governor's Chief of

Staff; the Director of the Texas Office of State-Federal

Relations; a pair of Texas Congressmen, and various other

interested parties, including engineers, physicists,

lawyers and one, other, lone journalist, from a Dallas

newspaper. Forty-six people in all, counting the Advisory

Committee of nine. Lunch was catered.

The mood was as dark as the fresh-drilled yet

already-decaying SSC tunnels. "I hope we can make

*something* positive out of all this," muttered US

Congressman Joe Barton (R-Tex), Waxahachie's

representative and a tireless champion of the original

project. A Texas state lawyer told me bitterly that "the

Department of Energy treats our wonderful asset like one

of their hazardous waste sites!"

For his part, the DoE's official representative, a

miserably unhappy flak-catcher from the Office of Energy

Research, talked a lot under extensive grilling by the

Committee, but said precisely nothing. "I honestly don't

know how the Secretary is going to write her report," he

mourned, wincing. "The policy is to close things down in

as cheap a way as possible."

Nothing about the SSC can be cleared without the nod

of the new Energy Secretary, the formidable Hazel O'Leary.

At the moment, Ms. O'Leary is very busy, checking the

DoE's back-files on decades of nuclear medical research on

uninformed American citizens. Her representative

conveyed the vague notion that Ms. O'Leary might be

inclined to allow something to be done with the site of

the SSC, if the State of Texas were willing to pay for

everything, and if it weren't too much trouble for her

agency. In the meantime she would like to cut the SSC's

shut-down budget for 1994 by two-thirds, with no money at

all for the SSC in 1995.

Hans Mark, former Chancellor of the University of

Texas System, gamely declared that the SSC would in fact

be built -- someday. Despite anything Congress may say,

the scientific need is still there, he told the committee

-- and Waxahachie is still the best site for such a

project. Mr. Mark compared the cancelled SSC to the

"cancelled" B-1 Bomber, a project that was built at last

despite the best efforts of President Carter to kill it.

"Five years down the road," he predicted, "or ten years."

He urged the State of Texas not to sell the 16,747 acres

it has purchased to house the site.

Federal engineering mandarin Bob White grimly called

the cancellation "a watershed in American science," noting

that never before had such a large project, of undisputed

scientific worth, been simply killed outright by Congress.

He noted that the physical assets of the SSC are worth

essentially nothing -- pennies per pound -- without the

trained staff, and that the staff is wasting away.

There remain some 1,983 people in the employ of the

SSC (or rather in the employ of the Universities Research

Association, a luckless academic bureaucracy that manages

the SSC and has taken most of the political blame for the

cost overruns). The dead Collider's technical staff

alone numbers over a thousand people: 16 in senior

management, 133 scientists, 56 applied physicists, 429

engineers, 159 computer specialists and network people,

159 guest scientists and research associates on grants

from other countries and other facilities, and 191

"technical associates."

"Deadwood," scoffed one attendee, "three hundred and

fifty people in physics research when we don't even have a

machine!" But the truth is that without a brilliantly

talented staff in place, all those one-of-a-kind cutting-

edge machines are so much junk. Many of those who stay

are staying in the forlorn hope of actually using some of

the smaller machines they have spent years developing and

building.

There have been, so far, about sixty more-or-less

serious suggestions for alternate uses of the SSC, its

facilities, its machineries, and its incomplete tunnel.

The SSC's Linear Accelerator was one of the smaller

assets of the great machine, but it is almost finished and

would be world-class anywhere else. It has been

repeatedly suggested that it could be used for medical

radiation treatments or for manufacturing medical

isotopes. Unfortunately, the Linear Accelerator is in

rural Ellis County, miles from Waxahachie and miles from

any hospital, and it was designed and optimized for

physics research, not for medical treatment or

manufacturing.

The former "N-15" site of the Collider, despite its

colorless name, is the most advanced manufacturing and

testing facility in the world -- when it comes to giant

superconducting magnets. The N-15 magnet facility is not

only well-nigh complete, but was almost entirely financed

by funds from the State of Texas. Unfortunately, the only

real market remaining for its "products" --

brobdingnagian frozen accelerator magnets -- is the

European CERN accelerator.

CERN itself has been hurting for money lately, its

German and Spanish government partners in particular

complaining loudly about the dire expense of hunting top

quarks and such.

Former SSC Director Roy Schwitters therefore declared

that CERN would need SSC's valuable magnets, and that the

US should use these assets as leverage for influence at

CERN.

This suggestion, however, was too much for Texan

Congressman Joe Barton. He described Schwitter's

suggestion as "very altruistic" and pointed out that the

Europeans had given the SSC "the back of their hand for

eight years!"

One could only admire the moral grit of SSC's former

Director in gamely proposing that the magnets, the very

backbone of his dead Collider, should be shipped, for the

good of science, to his triumphant European rivals. It

would seem that the American particle-physics research has

suffered such a blow from the collapse of the SSC that the

only reasonable course of action for the American physics

community is to go cap in hand to the Europeans and try,

somehow, to make things up.

At least, that proposal, galling as it may be, does

make some sense for American physicists -- but for an

American politician, to drop two billion dollars on the

SSC just to ship its magnets to some cyclotron in

Switzerland is quite another matter. When an attendee

gently urged Congressman Barton to "take a longer view" -

- perhaps, someday, the Europeans would reciprocate the

scientific favor -- the Texan Congressman merely narrowed

his eyes in a glare that would have scared Clint Eastwood,

and vowed "I will 'reciprocate' the concern that the

Europeans have shown for the SSC!"

It's been suggested that the numerous well-appointed

SSC offices could become campuses of some new research

institution: on magnets, or cryogenics, or controls, or

computer simulation. The physics departments of many

Texas colleges and universities like this idea. After

all, there's a great deal of handy state-of-the-art

clutter there, equipment any research lab in the world

would envy. Six and a half million dollars' worth of

machine tools and welding equipment. Three million in

high-tech calibration equipment and measuring devices.

Ten million dollars in trucks, vans, excavators,

bulldozers and such. A million-dollar print shop.

And almost fifty million dollars worth of state-of-

the-art computing equipment circa 1991 or so, including a

massively parallel Hypercube simulator, CAD/CAM

engineering and design facilities with millions of man-

hours of custom software, FDDI, OSI, and videoconferencing

office computer networks, and 2,600 Macintosh IIvx

personal computers. Plus a two-million dollar, fully-

equipped physics library.

Unfortunately it's very difficult to propose a new

physics facility just to make use of this, well, stuff,

when there are long-established federal physics research

facilities such as Los Alamos and Lawrence Livermore, now

going begging because nobody wants their veteran personnel

to build new nuclear weapons. If anyone builds such a

place in Waxahachie, then the State of Texas will have to

pay for it. And Texas is not inclined to shell out more

money. Texas already feels that the rest of the United

States owes Texas $440,885,853 for the dead Collider.

Besides the suggestions for medical uses, magnetic and

superconductive studies, and the creation of some new

research institute, there are the many suggestions

collectively known as "Other." One is to privatize the

SSC as the "American Institute for Superconductivity

Competitiveness" and ask for corporate help.

Unfortunately the hottest (or maybe "coolest") research

area in superconductivity these days is not giant helium-

frozen magnets for physicists, but the new ceramic

superconductors.

Other and odder schemes include a compressed-air

energy-storage research facility. An earth-wobble

geophysics experiment. Natural gas storage.

And, perhaps inevitably, the suggestion of Committee

member Martin Goland that the SSC tunnel be made into a

high-level nuclear waste-storage site. A "temporary"

waste site, he assured the Committee, that would store

highly radioactive nuclear waste in specially designed

"totally safe" steel shipping casks, until a "permanent"

site opens somewhere in New Mexico.

"I'm gonna sell my house now," stage-whispered the

physicist next to me in the audience. "Waxahachie will be

a ghost town!"

This was an upshot worthy of Greek myth -- a tunnel

built to steal the fiery secrets of the God Particle,

which ends up constipated by thousands of radioactive

steel coprolites, the Trojan Horse gift of Our Friend Mr.

Atom. It's such a darkly poetic, Southern-Gothic

example of hubris clobbered by nemesis that one almost

wishes it would actually happen.

As far as safety goes, hiding nuclear waste in an

incomplete 14.7 mile tunnel under Texas is certainly far

more safe than leaving the waste where it is at the moment

(basically, all over America, from sea to shining sea).

DoE's nuclear-waste chickens have come back to roost in

major fashion lately, as time catches up with a generation

of Cold War weapons scientists. "They were never given

the money they needed to do it cleanly, but just told to

do it right away in the name of National Security," a

federal expert remarked glumly over the ham and turkey

sandwiches at the lunch break. He went on to grimly

mention "huge amounts of carbon tetrachloride seeping into

the water table" and radioactive waste "storage tanks

that burp hydrogen."

But the Texans were having none of that; the chairman

of the Committee declared that they had heard Mr.

Goland's suggestion, and that it would go no further. The

room erupted into nervous laughter.

The Committee's first meeting broke up with the

suggestion that sixty million dollars be found somewhere-

or-other to maintain an unspecified "core staff" of SSC

researchers, while further study is undertaken on what to

actually do with the remains.

As the head of SMU's physics department has remarked,

"The general impression was that it would be an

embarrassment or a waste or sinful to say that, after $2

billion, you get nothing, zip, zero for it." However,

zip and zero may well be exactly the result, despite the

best intentions of the Texan clean-up crew. The dead

Collider is a political untouchable now. The Texans would

like to make something from the corpse, not for its own

sake, really, but just so the people of Texas will not

look quite so much like total hicks and chumps. The DoE,

for its part, would like this relic of nutty Reagan

Republicanism to vanish into the memory hole with all

appropriate speed. The result is quite likely to be a

lawsuit by the State of Texas against the DoE, where yet

more millions are squandered in years of wrangling by

lawyers, an American priesthood whose voracious appetite

for public funds puts even physicists to shame.

But perhaps "squandered" is too harsh a word for the

SSC. After all, it's not as if those two billion dollars

were actually spent on the subatomic level. They were

spent in perfectly normal ways, and went quite legally

into the pockets of standard government contractors such

as Sverdrup and EG&G (facilities construction), Lockheed

(systems engineering), General Dynamics, Westinghouse,

and Babcock and Wilcox (magnets), Obayashi & Dillingham

(tunnel contractors), and Robbins Company (Tunnel Boring

Machine). The money went to architects and engineers and

designers and roadpavers and people who string Ethernet

cable and sell UNIX boxes and Macintoshes. Those dollars

also paid the salaries of 2,000 researchers for several

years. Admittedly, the nation would have been far

better off it those 2,000 talented people simply had been

given a million dollars each and told to go turn

themselves into anything except particle physicists, but

that option wasn't presented.

The easy-going town of Waxahachie seems to have few

real grudges over the experience. A public meeting,

called so that sufferers in Waxahachie could air their

economic complaints about the dead Collider, had almost no

attendees. The entire bizarre enterprise seems scarcely

to have impinged at all on everyday life in Waxahachie.

Besides, not five miles from the SSC's major campus,

the Waxahachians still have their "Scarborough Fair," a

huge mock-medieval "English Village" where drawling "lords

and ladies" down on day-trips from Dallas can watch fake

jousts and drink mead in a romantic heroic-fantasy

atmosphere with ten times the popular appeal of that

tiresome hard-science nonsense.

As boondoggles go, SSC wasn't small. However, SSC

wasn't anywhere near so grotesque as the multiple billions

spent, both openly and covertly, on American military

science funding. Many of the SSC's contractors were in

fact military-industrial contractors, and it may have done

them some good to find (slightly) alternate employment.

The same goes for the many Russian nuclear physicists

employed by the SSC, who earned useful hard currency and

were spared the grim career-choices in Russia's collapsing

nuclear physics enterprise. It has been a cause of some

concern lately that Russian nuclear physicists may, as

Lederman and Glashow once put it, "go on to play vital

roles in the rest of the world" -- i.e., in the nuclear

enterprises of Libya, North Korea, Syria and Iraq. It's a

pity those Russians can't be put to work salting the tails

of quarks inside the SSC; though a cynic might say it's a

greater pity that they were ever taught physics in the

first place.

SCIENCE magazine, in its editorial post-mortem "The

Lessons of the Super Collider," had its own morals to

draw. Lesson One: "High energy physics has become too

expensive to be defined by national boundaries." Lesson

Two: "Just because particle physics asks questions about

the fundamental structure of matter does not give it any

greater claim on taxpayer dollars than solid-state physics

or molecular biology. Proponents of any project must

justify the costs in relation to the scientific and social

return."

That may indeed be the New Reality for American

science funding today, but it was never the justification

of the Machine in the Desert. The Machine in the Desert

was an absolute vision, about the absolute need to know.

And it was about pride. "Pride," wrote Lederman and

Glashow in 1985, "is one of the seven deadly sins," yet

they nevertheless declared their pride in the successes of

their predecessors, and their unbounded determination to

make America not merely the best in particle physics, but

the best in everything, as America had been when they were

children.

In his own 1993 post-mortem on the dead Collider,

written for the New York Times, Lederman raised the

rhetorical question, "Is the real problem the hubris of

physicists to believe that society would continue to

support this exploration no matter what the cost?" A

rhetorical question because Lederman, having raised that

cogent question, never bothered to address it. Instead,

he ended his column by blaming the always-convenient

spectre of American public ignorance of science. "Most

important of all," he concluded, "scientists must

rededicate themselves to a massive effort at raising the

science literacy of the general public. Only when the

citizens have a reasonable science savvy will their

congressional servants vote correctly."

Alas, many of our congressional servants already

possess plenty of science savvy; what they have, is

science savvy to their own ends. Not science for the sake

of Galileo, Newton, Maxwell, Einstein or Leon Lederman,

but science for the sake of the devil's bargain American

science has made with its political sponsors: knowledge

as power.

As for the supposedly ignorant general public, the

American public were far more generous with scientists

when scientists were very few in number, and regarded with

a proper superstitious awe by a mainly agricultural and

blue-collar populace. The more they come to understand

science, the less respect the American general public has

for the whims of its practitioners. Americans may not do

a lot of calculus, but most American voters are "knowledge

workers" of one sort or another nowadays, and they've

seen Carl Sagan on TV often enough to know that, even

though Carl's a nice guy, billions of stars and zillions

of quarks won't put bread on their tables. Raising the

general science literacy of the American public is

probably a self-defeating effort when it comes to monster

projects like the SSC. Teaching more American kids more

math and science will only increase the already vast

armies of scientists and federally funded researchers,

drastically shrinking the pool of available funds

tomorrow.

It's an open question whether a 40TeV collider like

the SSC will ever be built, by anyone, anywhere, ever.

The Europeans, in their low-key, suave, yet subtly

menacing fashion, seem confident that they can snag the

Higgs scalar boson with their upgraded CERN collider at a

mere tenth of the cost of Reagan's SSC. If so, corks

will pop in Zurich and there will be gnashing of teeth in

Brookhaven and Berkeley. American scientific competitors

will taste some of the agony of intellectual defeat in the

realm of physics that European scientists have been

swallowing yearly since 1945. That won't mean the end of

the world.

On the other hand, the collapse of SSC may well suck

CERN down in the backdraft. It may be that the global

prestige of particle physics has now collapsed so utterly

that European governments will also stop signing the

checks, and CERN itself will fail to build its upgrade.

Or even if they do build it, they may be simply

unlucky, and at 10 TeV the CERN people may get little to

show.

In which case, it may be that the entire pursuit of

particle physics, stymied by energy limits, will simply go

out of intellectual fashion. If the global revulsion

against both nuclear weapons and nuclear power increases

and intensifies, it is not beyond imagination to imagine

nuclear research simply dwindling away entirely. The

whole kit-and-caboodle of pions, mesons, gluinos,

antineutrinos, that whole strange charm of quarkiness, may

come to seem a very twentieth-century enthusiasm.

Something like the medieval scholastic enthusiasm for

numbering the angels that can dance on the head of a pin.

Nowadays that's a byword for a silly waste of intellectual

effort, but in medieval times that was actually the very

same inquiry as modern particle physics: a question about

the absolute limits of space and material being.

Or the SSC may never be built for entirely different

reasons. It may be that accelerating particles in the

next century will not require the massive Rube Goldberg

apparatus of a fifty-four-mile tunnel and the twelve

cryogenic plants with their entire tank farms of liquid

helium. It is a bit hard to believe that scientific

questions as basic as the primal nature of matter will be

abandoned entirely, but there is more than one way to

boost a particle. Giant *room-temperature*

superconductors really would transform the industrial

base, and they might make quarks jump hoops without the

macho necessity of being "super" at all.

In the end, it is hard to wax wroth at the dead

Collider, its authors, or those who pulled the plug. The

SSC was both sleazy and noble: at one level a "quark-

barrel" commercialized morass of contractors scrambling at

the federal trough, while Congressmen eye-gouged one

another in the cloakroom, scientists angled for the main

chance and a steady paycheck, and supposedly dignified

scholars ground their teeth in public and backbit like a

bunch of jealous prima donnas. And yet at the same time,

the SSC really was a Great Enterprise, a scheme to gladden

the heart of Democritus and Newton and Tycho Brahe, and

all those other guys who had no real job or a fat state

sinecure.

The Machine in the Desert was a transcendant scheme to

steal cosmic secrets, an enterprise whose unashamed raison

d'etre was to enable wild and glorious flights of

imagination and comprehension. It was sense-of-wonder

and utter sleaze at one and the same time. Rather like

science fiction, actually. Not that the SSC itself was

science fictional, although it certainly was (and is). I

mean, rather, that the SSC was very like the actual

writing and publishing of science fiction, an enterprise

where bright but surprisingly naive people smash galaxies

for seven cents a word and a chance at a plastic brick.

It would take a hard-hearted science fiction writer

indeed to stand at the massive lip of that 240-foot hole

in the ground at N15 -- as I did late one evening in

January, with the sun at my back and tons of hardware

gently rusting all around me and not a human being in

sight -- and not feel a deep sense of wonder and pity.

In another of his determined attempts to enlighten the

ignorant public, in his book THE GOD PARTICLE, Leon

Lederman may have said it best.

In a parody of the Bible called "The Very New

Testament," he wrote:

"And it came to pass, as they journeyed from the east,

that they found a plain in the land of Waxahachie, and

they dwelt there. And they said to one another, Go to,

let us build a Giant Collider, whose collisions may reach

back to the beginning of time. And they had

superconducting magnets for bending, and protons had they

for smashing.

"And the Lord came down to see the accelerator, which

the children of men builded. And the Lord said, Behold

the people are unconfounding my confounding. And the Lord

sighed and said, Go to, let us go down, and there give

them the God Particle so that they may see how beautiful

is the universe I have made."

A man who justifies his own dreams in terms of

frustrating God and rebuilding the Tower of Babel -- only

this time in Texas, and this time done right -- has got to

be utterly tone-deaf to his own intellectual arrogance.

Worse yet, the Biblical parody is openly blasphemous,

unnecessarily alienating a large section of Lederman's

potential audience of American voters. Small wonder that

the scheme came to grief -- great wonder, in fact, that

Lederman's Babel came anywhere as near to success as it

did.

Nevertheless, I rather like the sound of that

rhetoric; I admire its sheer cosmic chutzpah. I

scarcely see what real harm has been done. (Especially

compared to the harm attendant on the works of Lederman's

colleagues such as Oppenheimer and Sakharov.) It's true

that a man was crushed to death building the SSC, but he

was a miner by profession, and mining is very hazardous

work under any circumstances. Two billion dollars was,

it's true, almost entirely wasted, but governments always

waste money, and after all, it was only money.

Give it a decade or two, to erase the extreme

humiliation naturally and healthfully attendant on this

utter scientific debacle. Then, if the United States

manages to work its way free of its fantastic burden of

fiscal irresponsibility without destroying the entire

global economy in the process, then I, for one, as an

American and Texan citizen, despite everything, would be

perfectly happy to see the next generation of particle

physicists voted another three billion dollars, and told

to get digging again.

Or even four billion dollars.

Okay, maybe five billion tops; but that's my final

offer.

"Bitter Resistance"

Two hundred thousand bacteria could easily lurk under the top half

of this semicolon; but for the sake of focussing on a subject that's too

often out of sight and out of mind, let's pretend otherwise. Let's pretend

that a bacterium is about the size of a railway tank car.

Now that our fellow creature the bacterium is no longer three

microns long, but big enough to crush us, we can get a firmer mental grip

on the problem at hand. The first thing we notice is that the bacterium is

wielding long, powerful whips that are corkscrewing at a blistering

12,000 RPM. When it's got room and a reason to move, the bacterium can

swim ten body-lengths every second. The human equivalent would be

sprinting at forty miles an hour.

The butt-ends of these spinning whips are firmly socketed inside

rotating, proton-powered, motor-hubs. It seems very unnatural for a

living creature to use rotating wheels as organs, but bacteria are serenely

untroubled by our parochial ideas of what is natural.

The bacterium, constantly chugging away with powerful interior

metabolic factories, is surrounded by a cloud of its own greasy spew. The

rotating spines, known as flagella, are firmly embedded in the bacterium's

outer hide, a slimy, lumpy, armored bark. Studying it closely (we evade

the whips and the cloud of mucus), we find the outer cell wall to be a

double-sided network of interlocking polymers, two regular, almost

crystalline layers of macromolecular chainmail, something like a tough

plastic wiffleball.

The netted armor, wrinkled into warps and bumps, is studded with

hundreds of busily sucking and spewing orifices. These are the

bacterium's "porins," pores made from wrapped-up protein membrane,

something like damp rolled-up newspapers that protrude through the

armor into the world outside.

On our scale of existence, it would be very hard to drink through a

waterlogged rolled-up newspaper, but in the tiny world of a bacterium,

osmosis is a powerful force. The osmotic pressure inside our bacterium

can reach 70 pounds per square inch, five times atmospheric pressure.

Under those circumstances, it makes a lot of sense to be shaped like a

tank car.

Our bacterium boasts strong, highly sophisticated electrochemical

pumps working through specialized fauceted porins that can slurp up and

spew out just the proper mix of materials. When it's running its osmotic

pumps in some nutritious broth of tasty filth, our tank car can pump

enough juice to double in size in a mere twenty minutes. And there's

more: because in that same twenty minutes, our bacterial tank car can

build in entire duplicate tank car from scratch.

Inside the outer wall of protective bark is a greasy space full of

chemically reactive goo. It's the periplasm. Periplasm is a treacherous

mess of bonding proteins and digestive enzymes, which can yank tasty

fragments of gunk right through the exterior hide, and break them up for

further assimilation, rather like chemical teeth. The periplasm also

features chemoreceptors, the bacterial equivalent of nostrils or taste-

buds.

Beneath the periplasmic goo is the interior cell membrane, a tender

and very lively place full of elaborate chemical scaffolding, where pump

and assembly-work goes on.

Inside the interior membrane is the cytoplasm, a rich ointment of

salts, sugars, vitamins, proteins, and fats, the tank car's refinery

treasure-house.

If our bacterium is lucky, it has some handy plasmids in its custody.

A plasmid is an alien DNA ring, a kind of fly-by-night genetic franchise

which sets up work in the midst of somebody else's sheltering cytoplasm.

If the bacterium is unlucky, it's afflicted with a bacteriophage, a virus

with the modus operandi of a plasmid but its own predatory agenda.

And the bacterium has its own native genetic material. Eukaryotic

cells -- we humans are made from eukaryotic cells -- possess a neatly

defined nucleus of DNA, firmly coated in a membrane shell. But bacteria

are prokaryotic cells, the oldest known form of life, and they have an

attitude toward their DNA that is, by our standards, shockingly

promiscuous. Bacterial DNA simply sprawls out amid the cytoplasmic

goo like a circular double-helix of snarled and knotted Slinkies.

Any plasmid or transposon wandering by with a pair of genetic

shears and a zipper is welcome to snip some data off or zip some data in,

and if the mutation doesn't work, well, that's just life. A bacterium

usually has 200,000 or so clone bacterial sisters around within the space

of a pencil dot, who are more than willing to take up the slack from any

failed experiment in genetic recombination. When you can clone yourself

every twenty minutes, shattering the expected laws of Darwinian heredity

merely adds spice to life.

Bacteria live anywhere damp. In water. In mud. In the air, as

spores and on dust specks. In melting snow, in boiling volcanic springs. In

the soil, in fantastic numbers. All over this planet's ecosystem, any liquid

with organic matter, or any solid foodstuff with a trace of damp in it,

anything not salted, mummified, pickled, poisoned, scorching hot or frozen

solid, will swarm with bacteria if exposed to air. Unprotected food

always spoils if it's left in the open. That's such a truism of our lives

that it may seem like a law of physics, something like gravity or entropy;

but it's no such thing, it's the relentless entrepreneurism of invisible

organisms, who don't have our best interests at heart.

Bacteria live on and inside human beings. They always have;

bacteria were already living on us long, long before our species became

human. They creep onto us in the first instants in which we are held to

our mother's breast. They live on us, and especially inside us, for as long

as we live. And when we die, then other bacteria do their living best to

recycle us.

An adult human being carries about a solid pound of commensal

bacteria in his or her body; about a hundred trillion of them. Humans have

a whole garden of specialized human-dwelling bacteria -- tank-car E. coli,

balloon-shaped staphylococcus, streptococcus, corynebacteria,

micrococcus, and so on. Normally, these lurkers do us little harm. On the

contrary, our normal human-dwelling bacteria run a kind of protection

racket, monopolizing the available nutrients and muscling out other rival

bacteria that might want to flourish at our expense in a ruder way.

But bacteria, even the bacteria that flourish inside us all our lives,

are not our friends. Bacteria are creatures of an order vastly different

from our own, a world far, far older than the world of multicellular

mammals. Bacteria are vast in numbers, and small, and fetid, and

profoundly unsympathetic.

So our tank car is whipping through its native ooze, shuddering from

the jerky molecular impacts of Brownian motion, hunting for a

chemotactic trail to some richer and filthier hunting ground, and

periodically peeling off copies of itself. It's an enormously fast-paced

and frenetic existence. Bacteria spend most of their time starving,

because if they are well fed, then they double in number every twenty

minutes, and this practice usually ensures a return to starvation in pretty

short order. There are not a lot of frills in the existence of bacteria.

Bacteria are extremely focussed on the job at hand. Bacteria make ants

look like slackers.

And so it went in the peculiar world of our acquaintance the tank

car, a world both primitive and highly sophisticated, both frenetic and

utterly primeval. Until an astonishing miracle occurred. The miracle of

"miracle drugs," antibiotics.

Sir Alexander Fleming discovered penicillin in 1928, and the power

of the sulfonamides was recognized by drug company researchers in 1935,

but antibiotics first came into general medical use in the 1940s and 50s.

The effects on the hidden world of bacteria were catastrophic. Bacteria

which had spent many contented millennia decimating the human race

were suddenly and swiftly decimated in return. The entire structure of

human mortality shifted radically, in a terrific attack on bacteria from

the world of organized intelligence.

At the beginning of this century, back in the pre-antibiotic year of

1900, four of the top ten leading causes of death in the United States

were bacterial. The most prominent were tuberculosis ("the white

plague," *Mycobacterium tuberculosis*) and pneumonia (*Streptococcus

pneumoniae,* *Pneumococcus*). The death rate in 1900 from

gastroenteritis (*Escherichia coli,* various *Campylobacter* species,

etc.) was higher than that for heart disease. The nation's number ten

cause of death was diphtheria (*Corynebacterium diphtheriae*). Bringing

up the bacterial van were gonorrhea, meningitis, septicemia, dysentery,

typhoid fever, whooping cough, and many more.

At the end of the century, all of these festering bacterial afflictions

(except pneumonia) had vanished from the top ten. They'd been replaced

by heart disease, cancer, stroke, and even relative luxuries of

postindustrial mortality, such as accidents, homicide and suicide. All

thanks to the miracle of antibiotics.

Penicillin in particular was a chemical superweapon of devastating

power. In the early heyday of penicillin, the merest trace of this

substance entering a cell would make the hapless bacterium literally

burst. This effect is known as "lysing."

Penicillin makes bacteria lyse because of a chemical structure

called "beta-lactam." Beta-lactam is a four-membered cyclic amide ring,

a molecular ring which bears a fatal resemblance to the chemical

mechanisms a bacterium uses to build its cell wall.

Bacterial cell walls are mostly made from peptidoglycan, a plastic-

like molecule chained together to form a tough, resilient network. A

bacterium is almost always growing, repairing damage, or reproducing,

so there are almost always raw spots in its cell wall that require

construction work.

It's a sophisticated process. First, fragments of not-yet-peptided

glycan are assembled inside the cytoplasm. Then the glycan chunks are

hauled out to the cell wall by a chemical scaffolding of lipid carrier

molecules, and they are fitted in place. Lastly, the peptidoglycan is

busily knitted together by catalyzing enzymes and set to cure.

But beta-lactam is a spanner in the knitting-works, which attacks

the enzyme which links chunks of peptidoglycan together. The result is

like building a wall of bricks without mortar; the unlinked chunks of

glycan break open under osmotic pressure, and the cell spews out its

innards catastrophically, and dies.

Gram-negative bacteria, of the tank-car sort we have been

describing, have a double cell wall, with an outer armor plus the inner cell

membrane, rather like a rubber tire with an inner tube. They can

sometimes survive a beta-lactam attack, if they don't leak to death. But

gram-positive bacteria are more lightly built and rely on a single wall

only, and for them a beta-lactam puncture is a swift kiss of death.

Beta-lactam can not only mimic, subvert and destroy the assembly

enzymes, but it can even eat away peptide-chain mortar already in place.

And since mammalian cells never use any peptidoglycans, they are never

ruptured by penicillin (although penicillin does sometimes provoke serious

allergic reactions in certain susceptible patients).

Pharmaceutical chemists rejoiced at this world-transforming

discovery, and they began busily tinkering with beta-lactam products,

discovering or producing all kinds of patentable, marketable, beta-lactam

variants. Today there are more than fifty different penicillins and

seventy-five cephalosporins, all of which use beta-lactam rings in one

form or another.

The enthusiastic search for new medical miracles turned up

substances that attack bacteria through even more clever methods.

Antibiotics were discovered that could break-up or jam-up a cell's protein

synthesis; drugs such as tetracycline, streptomycin, gentamicin, and

chloramphenicol. These drugs creep through the porins deep inside the

cytoplasm and lock onto the various vulnerable sites in the RNA protein

factories. This RNA sabotage brings the cell's basic metabolism to a

seething halt, and the bacterium chokes and dies.

The final major method of antibiotic attack was an assault on

bacterial DNA. These compounds, such as the sulphonamides, the

quinolones, and the diaminopyrimidines, would gum up bacterial DNA

itself, or break its strands, or destroy the template mechanism that reads

from the DNA and helps to replicate it. Or, they could ruin the DNA's

nucleotide raw materials before those nucleotides could be plugged into

the genetic code. Attacking bacterial DNA itself was the most

sophisticated attack yet on bacteria, but unfortunately these DNA

attackers often tended to be toxic to mammalian cells as well. So they

saw less use. Besides, they were expensive.

In the war between species, humanity had found a full and varied

arsenal. Antibiotics could break open cell walls, choke off the life-giving

flow of proteins, and even smash or poison bacterial DNA, the central

command and control center. Victory was swift, its permanence seemed

assured, and the command of human intellect over the realm of brainless

germs was taken for granted. The world of bacteria had become a

commercial empire for exploitation by the clever mammals.

Antibiotic production, marketing and consumption soared steadily.

Nowadays, about a hundred thousand tons of antibiotics are

manufactured globally every year. It's a five billion dollar market.

Antibiotics are cheap, far cheaper than time-consuming, labor-intensive

hygienic cleanliness. In many countries, these miracle drugs are routinely

retailed in job-lots as over-the-counter megadosage nostrums.

Nor have humans been the only mammals to benefit. For decades,

antibiotics have been routinely fed to American livestock. Antibiotics

are routinely added to the chow in vast cattle feedlots, and antibiotics are

fed to pigs, even chickens. This practice goes on because a meat animal

on antibiotics will put on poundage faster, and stay healthier, and supply

the market with cheaper meat. Repeated protests at this practice by

American health authorities have been successfully evaded in courts and

in Congress by drug manufacturers and agro-business interests.

The runoff of tainted feedlot manure, containing millions of pounds

of diluted antibiotics, enters rivers and watersheds where the world's

free bacteria dwell.

In cities, municipal sewage systems are giant petri-dishes of

diluted antibiotics and human-dwelling bacteria.

Bacteria are restless. They will try again, every twenty minutes.

And they never sleep.

Experts were aware in the 1940s and 1950s that bacteria could, and

would, mutate in response to selection pressure, just like other

organisms. And they knew that bacteria went through many generations

very rapidly, and that bacteria were very fecund. But it seemed that any

bacteria would be very lucky to mutate to successfully resist even one

antibiotic. Compounding that luck by evolving to resist two antibiotics at

once seemed well-nigh impossible. Bacteria were at our mercy. They

didn't seem any more likely to resist penicillin and tetracycline than a

rainforest can resist bulldozers and chainsaws.

However, thanks to convenience and the profit motive, once-

miraculous antibiotics had become a daily commonplace. A general

chemical haze of antibiotic pollution spread across the planet. Titanic

numbers of bacteria, in all niches of bacterial life, were being given an

enormous number of chances to survive antibiotics.

Worse yet, bacteriologists were simply wrong about the way that

bacteria respond to a challenge.

Bacteria will try anything. Bacteria don't draw hard and fast

intellectual distinctions between their own DNA, a partner's DNA, DNA

from another species, virus DNA, plasmid DNA, and food.

This property of bacteria is very alien to the human experience. If

your lungs were damaged from smoking, and you asked your dog for a

spare lung, and your dog, in friendly fashion, coughed up a lung and gave

it to you, that would be quite an unlikely event. It would be even more

miraculous if you could swallow a dog's lung and then breathe with it just

fine, while your dog calmly grew himself a new one. But in the world of

bacteria this kind of miracle is a commonplace.

Bacteria share enormous amounts of DNA. They not only share

DNA among members of their own species, through conjugation and

transduction, but they will encode DNA in plasmids and transposons and

packet-mail it to other species. They can even find loose DNA lying

around from the burst bodies of other bacteria, and they can eat that DNA

like food and then make it work like information. Pieces of stray DNA can

be swept all willy-nilly into the molecular syringes of viruses, and

injected randomly into other bacteria. This fetid orgy isn't what Gregor

Mendel had in mind when he was discovering the roots of classical genetic

inheritance in peas, but bacteria aren't peas, and don't work like peas, and

never have. Bacteria do extremely strange and highly inventive things

with DNA, and if we don't understand or sympathize, that's not their

problem, it's ours.

Some of the best and cleverest information-traders are some of the

worst and most noxious bacteria. Such as *Staphylococcus *(boils).

*Haemophilus* (ear infections). *Neisseria *(gonorrhea).

Pseudomonas (abcesses, surgical infections). Even *Escherichia,* a very

common human commensal bacterium.

When it comes to resisting antibiotics, bacteria are all in the effort

together. That's because antibiotics make no distinctions in the world of

bacteria. They kill, or try to kill, every bacterium they touch.

If you swallow an antibiotic for an ear infection, the effects are not

confined to the tiny minority of toxic bacteria that happen to be inside

your ear. Every bacterium in your body is assaulted, all hundred trillion

of them. The toughest will not only survive, but they will carefully store,

and sometimes widely distribute, the genetic information that allowed

them to live.

The resistance from bacteria, like the attack of antibiotics, is a

multi-pronged and sophisticated effort. It begins outside the cell, where

certain bacteria have learned to spew defensive enzymes into the cloud of

slime that surrounds them -- enzymes called beta-lactamases,

specifically adapted to destroy beta-lactam, and render penicillin useless.

At the cell-wall itself, bacteria have evolved walls that are tougher and

thicker, less likely to soak up drugs. Other bacteria have lost certain

vulnerable porins, or have changed the shape of their porins so that

antibiotics will be excluded instead of inhaled.

Inside the wall of the tank car, the resistance continues. Bacteria

make permanent stores of beta-lactamases in the outer goo of periplasm,

which will chew the drugs up and digest them before they ever reach the

vulnerable core of the cell. Other enzymes have evolved that will crack

or chemically smother other antibiotics.

In the pump-factories of the inner cell membrane, new pumps have

evolved that specifically latch on to antibiotics and spew them back out

of the cell before they can kill. Other bacteria have mutated their interior

protein factories so that the assembly-line no longer offers any sabotage-

sites for site-specific protein-busting antibiotics. Yet another strategy

is to build excess production capacity, so that instead of two or three

assembly lines for protein, a mutant cell will have ten or fifty, requiring

ten or fifty times as much drug for the same effect. Other bacteria have

come up with immunity proteins that will lock-on to antibiotics and make

them useless inert lumps.

Sometimes -- rarely -- a cell will come up with a useful mutation

entirely on its own. The theorists of forty years ago were right when they

thought that defensive mutations would be uncommon. But spontaneous

mutation is not the core of the resistance at all. Far more often, a

bacterium is simply let in on the secret through the exchange of genetic

data.

Beta-lactam is produced in nature by certain molds and fungi; it was

not invented from scratch by humans, but discovered in a petri dish. Beta-

lactam is old, and it would seem likely that beta-lactamases are also very

old.

Bacteriologists have studied only a few percent of the many

microbes in nature. Even those bacteria that have been studied are by no

means well understood. Antibiotic resistance genes may well be present

in any number of different species, waiting only for selection pressure to

manifest themselves and spread through the gene-pool.

If penicillin is sprayed across the biosphere, then mass death of

bacteria will result. But any bug that is resistant to penicillin will

swiftly multiply by millions of times, thriving enormously in the power-

vacuum caused by the slaughter. The genes that gave the lucky winner its

resistance will also increase by millions of times, becoming far more

generally available. And there's worse: because often the resistance is

carried by plasmids, and one single bacterium can contain as many as a

thousand plasmids, and produce them and spread them at will.

That genetic knowledge, once spread, will likely stay around a while.

Bacteria don't die of old age. Bacteria aren't mortal in the sense that we

understand mortality. Unless they are killed, bacteria just keep splitting

and doubling. The same bacterial "individual" can spew copies of itself

every twenty minutes, basically forever. After billions of generations,

and trillions of variants, there are still likely to be a few random

oldtimers around identical to ancestors from some much earlier epoch.

Furthermore, spores of bacteria can remain dormant for centuries, then

sprout in seconds and carry on as if nothing had happened. This gives the

bacterial gene-pool -- better described as an entire gene-ocean -- an

enormous depth and range. It's as if Eohippus could suddenly show up at

the Kentucky Derby -- and win.

It seems likely that many of the mechanisms of bacterial resistance

were borrowed or kidnapped from bacteria that themselves produce

antibiotics. The genus Streptomyces, which are filamentous, Gram-

positive bacteria, are ubiquitous in the soil; in fact the characteristic

"earthy" smell of fresh soil comes from Streptomyces' metabolic products.

And Streptomyces bacteria produce a host of antibiotics, including

streptomycin, tetracycline, neomycin, chloramphenicol, and erythromycin.

Human beings have been using streptomycin's antibiotic poisons

against tuberculosis, gonorrhea, rickettsia, chlamydia, and candida yeast

infection, with marked success. But in doing so, we have turned a small-

scale natural process into a massive industrial one.

Streptomyces already has the secret of surviving its own poisons.

So, presumably, do at least some of streptomyces's neighbors. If the

poison is suddenly broadcast everywhere, through every niche in the

biosphere, then word of how to survive it will also get around.

And when the gospel of resistance gets around, it doesn't come just

one chapter at a time. Scarily, it tends to come in entire libraries. A

resistance plasmid (familiarly known to researchers as "R-plasmids,"

because they've become so common) doesn't have to specialize in just one

antibiotic. There's plenty of room inside a ring of plasmid DNA for handy

info on a lot of different products and processes. Moving data on and off

the plasmid is not particularly difficult. Bacterial scissors-and-zippers

units known as "transposons" can knit plasmid DNA right into the central

cell DNA -- or they can transpose new knowledge onto a plasmid. These

segments of loose DNA are aptly known as "cassettes."

So when a bacterium is under assault by an antibiotic, and it

acquires a resistance plasmid from who-knows where, it can suddenly

find an entire arsenal of cassettes in its possession. Not just resistance

to the one antibiotic that provoked the response, but a whole Bible of

resistance to all the antibiotics lately seen in the local microworld.

Even more unsettling news has turned up in a lab report in the

Journal of Bacteriology from 1993. Tetracycline-resistant strains in the

bacterium Bacteroides have developed a kind of tetracycline reflex.

Whenever tetracycline appears in the neighborhood, a Bacteroides

transposon goes into overdrive, manufacturing R-plasmids at a frantic

rate and then passing them to other bacteria in an orgy of sexual

encounters a hundred times more frequent than normal. In other words,

tetracycline itself now directly causes the organized transfer of

resistance to tetracycline. As Canadian microbiologist Julian Davies

commented in Science magazine (15 April 1994), "The extent and

biochemical nature of this phenomenon is not well understood. A number

of different antibiotics have been shown to promote plasmid transfer

between different bacteria, and it might even be considered that some

antibiotics are bacterial pheromones."

If this is the case, then our most potent chemical weapons have been

changed by our lethal enemies into sexual aphrodisiacs.

The greatest battlegrounds of antibiotic warfare today are

hospitals. The human race is no longer winning. Increasingly, to enter a

hospital can make people sick. This is known as "nosocomial infection,"

from the Latin for hospital. About five percent of patients who enter

hospitals nowadays pick up an infection from inside the hospital itself.

An epidemic of acquired immune deficiency has come at a

particularly bad time, since patients without natural immunity are forced

to rely heavily on megadosages of antibiotics. These patients come to

serve as reservoirs for various highly resistant infections. So do patients

whose immune systems have been artificially repressed for organ

transplantion. The patients are just one aspect of the problem, though;

healthy doctors and nurses show no symptoms, but they can carry strains

of hospital superbug from bed to bed on their hands, deep in the pores of

their skin, and in their nasal passages. Superbugs show up in food, fruit

juices, bedsheets, even in bottles and buckets of antiseptics.

The advent of antibiotics made elaborate surgical procedures safe

and cheap; but nowadays half of nosocomial infections are either surgical

infections, or urinary tract infections from contaminated catheters.

Bacteria are attacking us where we are weakest and most vulnerable, and

where their own populations are the toughest and most battle-hardened.

From hospitals, resistant superbugs travel to old-age homes and day-care

centers, predating on the old and the very young.

*Staphylococcus aureus,* a common hospital superbug which

causes boils and ear infections, is now present in super-strains highly

resistant to every known antibiotic except vancomycin. Enterococcus is

resistant to vancomycin, and it has been known to swap genes with

staphylococcus. If staphylococcus gets hold of this resistance

information, then staph could become the first bacterial superhero of the

post-antibiotic era, and human physicians of the twenty-first century

would be every bit as helpless before it as were physicians of the 19th. In

the 19th century physicians dealt with septic infection by cutting away

the diseased flesh and hoping for the best.

Staphylococcus often lurks harmlessly in the nose and throat.

*Staphylococcus epidermis,* a species which lives naturally on human

skin, rarely causes any harm, but it too must battle for its life when

confronted with antibiotics. This harmless species may serve as a

reservoir of DNA data for the bacterial resistance of other, truly lethal

bacteria. Certain species of staph cause boils, others impetigo. Staph

attacking a weakened immune system can kill, attacking the lungs

(pneumonia) and brain (meningitis). Staph is thought to cause toxic shock

syndrome in women, and toxic shock in post-surgical patients.

A 1994 outbreak of an especially virulent strain of the common

bacterium Streptococcus, "necrotizing fasciitis," caused panic headlines

in Britain about "flesh-eating germs" and "killer bugs." Of the fifteen

reported victims so far, thirteen have died.

A great deal has changed since the 1940s and 1950s. Strains of

bacteria can cross the planet with the speed of jet travel, and populations

of humans -- each with their hundred trillion bacterial passengers --

mingle as never before. Old-fashioned public-health surveillance

programs, which used to closely study any outbreak of bacterial disease,

have been dismantled, or put in abeyance, or are underfunded. The

seeming triumph of antibiotics has made us careless about the restive

conquered population of bacteria.

Drug companies treat the standard antibiotics as cash cows, while

their best-funded research efforts currently go into antiviral and

antifungal compounds. Drug companies follow the logic of the market;

with hundreds of antibiotics already cheaply available, it makes little

commercial sense to spend millions developing yet another one. And the

market is not yet demanding entirely new antibiotics, because the

resistance has not quite broken out into full-scale biological warfare.

And drug research is expensive and risky. A hundred million dollars of

investment in antibiotics can be wiped out by a single point-mutation in a

resistant bacterium.

We did manage to kill off the smallpox virus, but none of humanity's

ancient bacterial enemies are extinct. They are all still out there, and

they all still kill people. Drug companies mind their cash flow, health

agencies become complaisant, people mind what they think is their own

business, but bacteria never give up. Bacteria have learned to chew up,

spit out, or shield themselves from any and every drug we can throw at

them. They can now defeat every technique we have. The only reason true

disaster hasn't broken out is because all bacteria can't all defeat all the

techniques all at once. Yet.

There have been no major conceptual breakthroughs lately in the

antibiotic field. There has been some encouraging technical news, with

new techniques such as rational drug design and computer-assisted

combinatorial chemistry. There may be entirely new miracle drugs just

over the horizon that will fling the enemy back once again, with enormous

losses. But on the other hand, there may well not be. We may already

have discovered all the best antibiotic tricks available, and squandered

them in a mere fifty years.

Anyway, now that the nature of their resistance is better

understood, no bacteriologist is betting that any new drug can foil our

ancient enemies for very long. Bacteria are better chemists than we are

and they don't get distracted.

If the resistance triumphs, it does not mean the outbreak of

universally lethal plagues or the end of the human race. It is not an

apocalyptic problem. What it would really mean -- probably -- is a slow

return, over decades, to the pre-antibiotic bacterial status-quo. A return

to the bacterial status-quo of the nineteenth century.

For us, the children of the miracle, this would mean a truly shocking

decline in life expectancy. Infant mortality would become very high; it

would once again be common for parents to have five children and lose

three. It would mean a return to epidemic flags, quarantine camps,

tubercular sanatariums, and leprosariums.

Cities without good sanitation -- mostly Third World cities --

would suffer from water-borne plagues such as cholera and dysentery.

Tuberculosis would lay waste the underclass around the world. If you cut

yourself at all badly, or ate spoiled food, there would be quite a good

chance that you would die. Childbirth would be a grave septic risk for the

mother.

The practice of medicine would be profoundly altered. Elaborate,

high-tech surgical procedures, such as transplants and prosthetic

implants, would become extremely risky. The expense of any kind of

surgery would soar, since preventing infection would be utterly necessary

but very tedious and difficult. A bad heart would be a bad heart for life,

and a shattered hip would be permanently disabling. Health-care budgets

would be consumed by antiseptic and hygienic programs.

Life without contagion and infection would seem as quaintly exotic

as free love in the age of AIDS. The decline in life expectancy would

become just another aspect of broadly diminishing cultural expectations

in society, economics, and the environment. Life in the developed world

would become rather pinched, wary, and nasty, while life in the

overcrowded human warrens of the megalopolitan Third World would

become an abattoir.

If this all seems gruesomely plausible, it's because that's the way

our ancestors used to live all the time. It's not a dystopian fantasy; it

was the miracle of antibiotics that was truly fantastic. It that miracle

died away, it would merely mean an entirely natural return to the normal

balance of power between humanity and our invisible predators.

At the close of this century, antibiotic resistance is one of the

gravest threats that confronts the human race. It ranks in scope with

overpopulation, nuclear disaster, destruction of the ozone, global

warming, species extinction and massive habitat destruction. Although it

gains very little attention in comparison to those other horrors, there is

nothing theoretical or speculative about antibiotic resistance. The mere

fact that we can't see it happening doesn't mean that it's not taking place.

It is occurring, stealthily and steadily, in a world which we polluted

drastically before we ever took the trouble to understand it.

We have spent billions to kill bacteria but mere millions to truly

comprehend them. In our arrogance, we have gravely underestimated our

enemy's power and resourcefulness. Antibiotic resistance is a very real

threat which is well documented and increasing at considerable speed. In

its scope and its depth and the potential pain and horror of its

implications, it may the greatest single menace that we human beings

confront -- besides, of course, the steady increase in our own numbers.

And if we don't somehow resolve our grave problems with bacteria, then

bacteria may well resolve that population problem for us.



Поделиться книгой:

На главную
Назад