by Peter Watson
Before Civilisation had two core arguments. First, it revised the timing of the way the earth was populated. For example, from about 1960 on it was known that Australia had been occupied by man as early as 4000 BC, and maybe even as early as 17000 BC. Maize, it was established, was gathered systematically in Mexico by 5000 BC, and well before 3000 BC it was showing signs of domestication. The real significance of these dates was not just that they were earlier than anyone had thought hitherto, but that they killed off the vague theories then current that Meso-America had only developed civilisation after it had been imported, in some indefinable way, from Europe. The Americas had been cut off from the rest of the world since 12000–13000 BC, in effect the last ice age, and had developed all the hallmarks of civilisation – farming, building, metallurgy, religion – entirely separately.46
This revision of chronology, and what it meant, was the second element in Renfrew’s book, and here he concentrated on the area he knew best, Europe and the classical world of the Middle East. In the traditional view, the civilisations of the Middle East – Sumer and Egypt, for example – were the mother civilisations, the first great collective achievements of mankind, giving rise to the Minoans on Crete, and the classical world of the Aegean: Athens, Mycenae, Troy. From there, civilisation had spread farther north, to the Balkans and then Germany and Britain, and west to Italy and then France, and the Iberian peninsula. But after the C14 revolution, there was suddenly a serious problem with this model.47 On the new dating, the huge megalithic sites of the Atlantic seaboard, in Spain and Portugal, in Brittany and Britain, and in Denmark, were either contemporaneous with the civilisations of the Aegean or actually preceded them. This wasn’t just a question of an isolated date here and there but of many hundreds of revised datings, consistent with each other, and which in some cases put the Atlantic megaliths up to a thousand years earlier than Aegean cultures. The traditional model, for Egypt, the Middle East, and the Aegean, still held. But there was, as Renfrew put it, a sort of archaeological ‘fault line’ around the Aegean. Beyond that, a new model was needed.
The model he came up with started from a rejection of the old idea of ‘diffusion’ – that there had been an area of mother civilisations in the Middle East from which ideas of farming, metallurgy, and, say, the domestication of plants and animals had started, and then spread to all other areas as people migrated. It seemed clear to Renfrew that up and down the Atlantic coasts of Europe, there had developed a series of chiefdoms, a level of social organisation midway between hunter-gatherers and full-blown civilisation as represented in Egypt, Sumer, or Crete, which had kings, elaborate palaces, a highly stratified society. The sovereign areas of the chiefdoms were smaller (six on the Isle of Arran in Scotland, for example), and they were centred around large tombs and occasionally religious/astronomical sites, such as Stonehenge.48 Associated with these chiefdoms were a rudimentary social stratification and early trade. Sufficient numbers were needed to build the impressive stone works, funerary religious monuments around which the clans cohered. The megaliths were always found associated with arable land, suggesting that chiefdoms were a natural stage in the evolution of society: when man settled with the first domesticated crops, chiefdoms and megaliths soon followed.49
Renfrew’s analysis, now generally accepted, concentrated on sites in Britain, Spain, and the Balkans, which illustrated his argument. But it was his general thrust that counted: although early man had no doubt spread out to populate the globe from an initial point (maybe East Africa), civilisation, culture – call it what you will – had not developed in one place and then spread in the same way; civilisations had grown up in different times at different places of their own accord.50 This had two important long-term intellectual consequences, quite apart from killing off any idea (which still lingered then) that American civilisations had been seeded by European ones (such as the ‘lost tribe’ of Israel). First, it demolished the idea that, culturally speaking, the history of man is one story; all cultures across the world were sui generis and did not owe their being to a mother culture, the ancestor of all. Combined with the findings of anthropologists, this made all cultures equally potent and equally original, and therefore the ‘classical’ world was no longer the ultimate source.
At a deeper level, as Renfrew specifically pointed out, the discoveries of the new archaeology showed the dangers of succumbing too easily to Darwinian thinking.51 The old diffusionist theory was a form of evolution, but a form of evolution so general as to be almost meaningless. It suggested that civilisations developed in an unbroken, single sequence. The new C14 and tree-ring evidence showed that that simply wasn’t true. The new view wasn’t any less ‘evolutionary,’ but it was very different. It was, above all, a cautionary tale.
32
HEAVEN AND EARTH
The words historic moment have been heavily overused in this century. But if any moment outside war can be described as truly historic it surely occurred at twenty seconds after 3:56 AM BST on Monday, 21 July 1969, when Neil Armstrong stepped off the ladder that led from the ‘Eagle,’ the landing module of the Apollo 11 spacecraft, and on to the surface of the Moon, making him the first person to arrive on a celestial body outside Earth. As he did so, he spoke the words that have since become famous: ‘That’s one small step for man – one giant leap for mankind.1
For the benefit of scientists back at Mission Control in Houston, he then went on in a more down-to-earth, scientifically informative way: ‘The surface is fine and powdery, I can … I can pick it up loosely with my toe. It does adhere in fine layers like powdered charcoal to the sole and sides of my boots. I can only go in a small fraction of an inch. Maybe an eighth of an inch, but I can see the footprints of my boots and the treads in the fine sandy particles…. There seems to be no difficulty in moving around, as we suspected…. We’re essentially on a level place here – very level place here.’2 If the greatest intellectual achievement of the first half of the twentieth century was undeniable – the conception and construction of the atomic bomb – the achievements of the second half were more diverse, including the isolation and understanding of DNA, and the computer. But space travel and the moon landing certainly count as one of the greatest achievements of the century.
After the Russians had sprung their surprise in 1957, stealing a march on the United States with the launch of Sputnik 1, they had built on their lead, putting the first animal in space, the first man (Yuri Gagarin, in 1961), and the first woman (Valentina Tereshkova, in 1963). The United States responded with something close to panic. President Kennedy called a frenzied meeting at the White House four days after the Bay of Pigs disaster (when 1,500 Cuban exiles, trained in America by the U.S. military, invaded the island, only to be killed or captured). Railing that ‘something must be done,’ he had shouted at Lyndon Johnson, the vice president, ordering him to find out if ‘we have a chance of beating the Soviets by putting a laboratory in space, or by a trip round the Moon, or by a rocket to land on the Moon, or by a rocket to go to the Moon and back with a man?’3 The Americans finally put John Glenn into orbit on 20 February 1962 (Alan Shephard made a fifteen-minute nonorbital flight in May 1961). But then the Americans began to catch up, thanks to Kennedy’s commitment to the Apollo program with its aim to land a manned spacecraft on the Moon ‘before this decade is out.’4 Begun in 1963 (though NASA had been created in 1958), America spent up to $5 billion a year on space over the next ten years. That sum gives an idea of the size of the project, which involved, among other things, building a reliable spaceship bigger than a railway engine, designing and manufacturing a rocket heavier than a destroyer, and inventing several new materials.5 The project benefited from the brains of 400,000 people from 150 universities and 20,000 firms. We already know from Korolev that rocket technology lay at the heart of the space program and the biggest U.S. rocket, Saturn 5, weighed 2,700 tons, roughly the same as 350 London buses. Developed under Wernher von Braun, another German emigré, Saturn was 364 feet high, had 2
million working parts, 2.5 million solder joints, and 41 separate engines for guidance purposes, and carried in all 11,400,000 gallons of fuel – liquid nitrogen, oxygen, hydrogen, and helium, some of it stored at minus 221 degrees centigrade to keep it liquid.6 The oxygen alone filled the equivalent of fifty-four railway container tanks.7 The Moonship contained the cone-shaped command module, the only part to come back to earth and which therefore needed to withstand the enormous temperatures on re-entry to the atmosphere (caused by friction at such high speeds).8 One of the main engineering problems was to keep the cryogenic fuels cool enough. The tanks eventually designed were so airtight that, were ice cubes to have been installed inside, they would not have melted for nine years. The exit hatch of the module needed 150 new tools to be invented. Some bolts had to be locked in place by two men using a five-foot wrench.9
No one really knew how conditions in space would affect the men.10 Great care was taken therefore with psychological selection and training. They were taught to be tolerant and careful (always avoiding acute angles where they might snag their suits), and they were given massages every day. The crews that advanced were those that had worked together in harmony for more than a year. Interestingly enough, over the years both the Americans and the Russians came up with a near-identical profile of the ideal astronaut: they should not be too old, no later than their late thirties, or too tall, no taller than five-eleven or six feet; they should be qualified jet and test pilots with degrees in engineering.11 Finally, there was the reconnaissance of the Moon itself. Quite apart from the prospect in time of colonising space and its minerals, there were sound scientific reasons for studying the Moon close up. Since it lacked an atmosphere, the Moon was in some senses in pristine condition, ‘a priceless antique,’ as one scientist called it, in much the same condition as it had been when the universe, or the solar system at least, had first evolved. Examination of the rocks would also help decide how the Moon formed – whether it was once part of Earth, or broke off with Earth from the sun after collision with an asteroid, or was formed by very hot gas cooling.12 Both American and Soviet space probes got progressively closer to the Moon, sending back better and better photographs until objects as small as five feet wide could be distinguished. Five areas were originally chosen for landing, then narrowed to one, the Sea of Tranquillity, actually a flat plain free of craters.13
The biggest disaster of the American program took place in 1967, when a spaceship caught fire on the launchpad at Cape Kennedy after liquid oxygen somehow ignited, killing all three men inside. The world never knew how many Russian astronauts perished because of the greater secrecy surrounding their program, but distress messages picked up by radio hams around the globe suggested that at least eight were in trouble between 1962 and 1967.14 The greatest drama before the moon landing itself was the December 1968 flight of Apollo 8 around the Moon, which involved going behind the Moon to the dark side, which no one had ever seen, and meant that the crew would be out of radio contact with Mission Control for about half an hour. If the ‘burn’ of the engines was too strong, it might veer off into deep space; if it was too weak, it might crash into the Moon, on the far side, never to be heard from again.15 The pope sent a message of goodwill, as did a number of Russian space scientists, acknowledging implicitly at this point that the Americans were now decisively ahead.
At 9:59 AM on Christmas Eve, Apollo 8 swung behind the Moon. Mission Control in Houston, and the rest of the world, waited. Ten minutes of silence passed; twenty; thirty. At 10:39 AM Frank Borman’s voice could be heard reporting data from his instruments. Apollo 8 was exactly on schedule and, as Peter Fairley narrates the episode in his history of the Apollo Project, after a journey of a quarter of a million miles, it had arrived on its trajectory within half a mile of the one planned.16
The scene was set for Apollo 11. Edward ‘Buzz’ Aldrin Jr. joined Neil Armstrong on the surface of the Moon, where they placed a plaque, and a flag, planted some seeds, and collected rock samples with specially designed tools that avoided them having to bend. Then it was back into the ‘Lunar Bug,’ rendezvous with Michael Collins in the command module, and the return journey, splashing down near Johnston Island in the Pacific, where they were met by the USS Hornet with President Richard Nixon on board. The men had returned safely to Earth, and the space age had begun.17
The landing on the Moon was, however, in some ways a climax rather than a debut. Crewed flights to the Moon continued until 1972, but then stopped. As the 1970s wore on, space probes went deeper into the heavens – Venus, Mars, Mercury, Jupiter, the Sun, Saturn, with Pioneer 10, launched in 1972, becoming the first manmade object to leave the solar system, which it did in 1983. Actual landings were considered less necessary after the first flush of excitement, and both the Americans and Russians concentrated on longer flights in orbit, to enable scientists to carry out experiments in space: in 1973, in the United States Skylab, astronauts spent eighty-four days aboard. The first stage of the space age may be said to have matured around 1980. In that year, Intelsat 5 was launched, capable of relaying thousands of telephone calls and two television channels. And in the following year the Columbia, the first reusable space shuttle, was launched. In just ten years space travel had gone from being exotic to being almost mundane.
*
The space race naturally stimulated interest in the heavens in general, a happy coincidence, as the 1960s had in any case seen some very important advances in our understanding of the universe, even without the advantages conferred by satellite technology. In the first half of the century, apart from the development of the atomic bomb and relativity, the main achievement of physics was its unification with chemistry (as epitomised in the work of Linus Pauling). After the war, the discovery of yet more fundamental particles, especially quarks, brought about an equivalent unification, between physics and astronomy. The result of this consilience, as it would be called, was a much more complete explanation of how the heavens – the universe – began and evolved. It was, for those who do not find the reference blasphemous, an alternative Genesis.
Quarks, as we have seen, were originally proposed by Murray Gell-Mann and George Zweig, almost simultaneously in 1962. It is important to grasp that quarks do not exist in isolation in nature (at least on Earth), but the significance of the quark (and certain other particles isolated in the 1960s and 1970s but which we need not describe here) is that it helps explain conditions in the early moments of the universe, just after the Big Bang. The idea that the universe began at a finite moment in the past was accepted by most physicists, and many others, since Hubble’s discovery of the red shift in 1929, but the 1960s saw renewed interest in the topic, partly as a result of Gell-Mann’s theories about the quark but also because of an accidental discovery made at the Bell Telephone Laboratories in New Jersey, in 1965.
Since 1964, the Bell Labs had been in possession of a new kind of telescope. An antenna located on Crawford Hill at Holmdel communicated with the skies via the Echo satellite. This meant the telescope was able to ‘see’ into space without the distorting interference of the atmosphere, and that far more of the skies were accessible. As their first experiment, the scientists in charge of the telescope, Arno Penzias and Robert Wilson, decided to study the radio waves being emitted from our own galaxy. This was essentially baseline research, the idea being that once they knew what pattern of radio waves we were emitting, it would be easier to study similar waves coming from elsewhere. Except that it wasn’t that simple. Wherever they looked in the sky, Penzias and Wilson found a persistent source of interference – like static. At first they had thought there was something wrong with their instruments. A pair of pigeons were nesting in the antenna, with the predictable result that there were droppings everywhere. The birds were captured and sent to another part of the Bell complex. They came back. This time, according to Steven Weinberg’s account published later, they were dealt with ‘by more decisive means.18 With the antenna cleaned up, the ‘static’ was reduced, but only
minimally, and it still appeared from all directions. Penzias discussed his mystery with another radio astronomer at MIT, Bernard Burke. Burke recalled that a colleague of his, Ken Turner of the Carnegie Institute of Technology, had mentioned a talk he had heard at Johns Hopkins University in Baltimore given by a young theorist from Princeton, P. J. E. Peebles, which might bear on the ‘static’ mystery. Peebles’s speciality was the early universe. This was a relatively new discipline and still very speculative. As we saw in chapter 29, in the 1940s an emigré from Ukraine, George Gamow, had begun to think about applying the new particle physics to the conditions that must have existed at the time of the Big Bang. He started with ‘primordial hydrogen,’ which, he said, would have been partly converted into helium, though the amount produced would have depended on the temperature of the Big Bang. He also said that the hot radiation corresponding to the enormous fireball would have thinned out and cooled as the universe expanded. He went on to argue that this radiation ‘should still exist, in a highly “red-shifted” form, as radio waves.19 This idea of ‘relict radiation’ was taken up by others, some of whom calculated that such radiation should now have a temperature of 5 K (i.e., 5 degrees above absolute zero). Curiously, with physics and astronomy only just beginning to come together, no physicist appeared to be aware that even then radio astronomy was far enough ahead to answer that question. So the experiment was never done. And when radio astronomers at Princeton, under Robert Dicke, began examining the skies for radiation, they never looked for the coolest kinds, not being aware of their significance. It was a classic case of the right hand not knowing what the left was doing. When Peebles, a Canadian from Winnipeg, started his Ph.D. at Princeton in the late 1950s, he worked under Robert Dicke. Gamow’s theories had been forgotten but, more to the point, Dicke himself seems to have forgotten his own earlier work.20 The result was that Peebles unknowingly repeated all the experiments and theorising of those who had gone before. He arrived at the same conclusion, that the universe should now be filled with ‘a sea of background radiation’ with a temperature of only a few K. Dicke, who either still failed to remember his earlier experiments or didn’t realise their significance, liked Peebles’s reasoning enough to suggest that they build a small radio telescope to look for the background radiation.