Book Read Free

Lawrence Krauss - The Greatest Story Ever Told--So Far

Page 3

by Why Are We Here (pdf)


  is no going back.

  ͣ͞

  C h a p t e r 2

  S E E I N G I N T H E D A R K

  Let there be light: and there was light.

  —GENESIS 1:3

  In the beginning there was light.

  It is no coincidence that the ancients imagined in Genesis that

  light was created on the first day. Without light, there would be little

  awareness of the vast universe surrounding us. When we nod and

  say, “I see,” to a friend who is trying to explain something, we convey

  far more than just an observation, but rather a fundamental

  understanding.

  Plato’s allegory was appropriately centered on light—light from a

  fire to cast the shadows on the cave wall and light from the outside

  to temporarily blind the freed prisoner and then illuminate the real

  world for him. Like the prisoners in the cave, we too are prisoners of

  light—almost everything we learn about the world we learn from

  what we see.

  While the most significant words in the Western religious canon

  may be Let there be light, in the modern world this phrase now has a

  completely different significance from what it once did. Human

  beings may be prisoners of light, but so is the universe. What once

  appeared as a whim of a Judeo-Christian God, or other gods before

  that one, we now understand to be required by the very laws that

  allow both heaven, and more important, Earth, to exist. You cannot

  have one without the other. Earth, or matter, follows light.

  ͤ͞

  This change in perception underlies almost every development in

  the edifice we call modern science. I am writing these words as I

  stare out from a ship at one of the Galápagos Islands, which Charles

  Darwin made famous, and which made him famous in return, as he

  changed our perception of life and its diversity with a single brilliant

  realization: that all living species developed through the natural

  selection of small inherited variations that are passed along to future

  generations by survivors. As surely as the understanding of evolution

  changed everything about our understanding of biology, our

  changing understanding of light changed everything about our

  physical understanding of our place in the universe. As a useful

  fringe benefit, this change resulted in virtually all of the technology

  on which the modern world is based.

  The extent to which our observations of the world imprison our

  minds, and frame our description of the fabric of the universe,

  remained unappreciated for more than twenty centuries following

  Plato. Once serious minds began to investigate in detail the hidden

  nature of the universe, it took over four centuries for them to fully

  resolve the question What is light?

  Perhaps the most serious modern mind, although certainly not

  the first, to ask this question was also one of the most famous—and

  oddest—scientists in history: Isaac Newton. It is not inappropriate to

  classify Newton as a modern mind—after all, his seventeenth-

  century Principia: Mathematical Principles of Natural Philosophy

  uncovered the classical laws of motion and laid the basis for his

  theory of gravity, both of which form the foundation of much of

  modern physics. Nevertheless, as John Maynard Keynes pointed out:

  Newton was not the first of the age of reason, he was the last of the

  magicians, the last of the Babylonians and Sumerians, the last

  great mind that looked out on the visible and intellectual world

  ͥ͞

  with the same eyes as those who began to build our intellectual

  inheritance rather less than 10,000 years ago.

  The truth of this statement reflects the revolutionary importance

  of Newton’s work. After the Principia, no rational person could view

  the world the same way the ancients had viewed it. But it also

  reflects the character of Newton himself. He devoted far more time,

  and far more ink, to writing about the occult, alchemy, and

  searching for hidden meanings and codes in the Bible—focusing in

  particular on the Book of Revelation and mysteries associated with

  the ancient Temple of Solomon—than he did to writing about

  physics.

  Newton was also one in a long line of people, which extends

  before and after him, who felt that he had been specifically chosen

  by God to help reveal the true meaning of the Scriptures. To what

  extent his studies of the universe derived from his fascination with

  the Bible is not clear, but it does seem reasonable to conclude that

  his primary interest was in theology, and that natural philosophy

  came in well below that, and probably below alchemy as well.

  Many individuals point to Newton’s fascination with God as

  evidence of the compatibility between science and religion, and to

  assert that modern science owes its existence to Christianity. This

  confuses history with causality. It is undeniable that many of the

  early giants of modern Western natural philosophy, from Newton

  onward, were deeply religious, although Darwin lost much, if not all,

  of his religious belief later in life. But remember that during much of

  this period there were primarily two sources of education and

  wealth: the Church and the Crown. The Church was the National

  Science Foundation of the fifteenth, sixteenth, and seventeenth

  centuries. All institutions of higher learning were tied to various

  denominations, and it was unthinkable for any educated person to

  not be affiliated with the Church. And as Giordano Bruno and later

  ͟͜

  Galileo discovered, it was unpleasant at best to counter its doctrine.

  It would have been remarkable for any of these leading early

  scientific thinkers to have been anything but religious.

  The religiosity of the early scientific pioneers is also cited today by

  sophists who claim that science and religious doctrine are

  compatible, but who confuse science and scientists. In spite of

  frequent appearances to the contrary, scientists are people. And like

  all people they are capable of holding many potentially mutually

  contradictory notions in their head at the same time. No correlation

  between divergent views held by any individual is representative of

  anything but human foibles.

  To claim that some scientists are or were religious is like saying

  some scientists are Republicans or some are flat-earthers or some are

  creationists. It doesn’t imply causality or consistency. My friend

  Richard Dawkins has told me of a professor of astrophysics who,

  during the day, writes papers that are published in astronomical

  journals assuming that the universe is more than 13 billion years old,

  but then goes home and privately espouses the literal biblical claim

  that the universe is six thousand years old.

  What determines intellectual consistency or lack thereof in the

  sciences is a combination of rational arguments with subsequent

  evidence and continued testing. It is perfectly reasonable to claim

  that religion, in the Western world, may be the mother of science.

  But as any parent knows, children rarely grow up to be
models of

  their parents.

  Newton may, following tradition, have been motivated to look at

  light because it was a gift from God. But we remember his work not

  because of his motivation, but because of what he discovered.

  Newton was convinced that light was made of particles, which he

  referred to as corpuscles, while Descartes, and later Newton’s

  nemesis Robert Hooke, and still later the Dutch scientist Christiaan

  ͟͝

  Huygens, all claimed that light was a wave. One of the key

  observations that appeared to support the wave theory was that

  white light, such as light from the Sun, could split into all the colors

  of the rainbow when passed through a prism.

  As was often the case during his life, Newton believed that he was

  correct and several of his most famous contemporaries (and

  competitors) were wrong. To demonstrate this, he devised a clever

  experiment using prisms that he first performed while at home in

  Woolsthorpe, to escape the bubonic plague ravaging Cambridge. As

  he reported at the Royal Society in 1672, on the forty-fourth try, he

  observed precisely what he hoped he would see.

  Advocates of the wave theory had argued that light waves were

  made of white light and that the light split into colors when it passed

  through a prism because of “corruption” of the rays as they traversed

  the glass. In this case, the more glass, the more splitting.

  Newton reasoned that this was not the case, but that light is made

  of colored particles that combine together to appear white. (With a

  nod to his occult fascination, Newton classified the colored particles

  of the spectrum—a term he coined—into seven different types: red,

  orange, yellow, green, blue, indigo, and violet. From the time of the

  Greeks, the number seven had been considered to possess mystical

  qualities.) To demonstrate that the wave/corruption picture was

  incorrect, Newton passed a beam of white light through two prisms

  held in opposite orientations. The first prism split the light into its

  spectrum, and the second recomposed it back into a single white

  light beam. This result would have been impossible if the glass had

  corrupted the light. A second prism would have simply made the

  situation worse and would not have caused the light to revert back to

  its original state.

  This result does not in fact disprove the wave theory of light (it

  actually supports it, because light slows down as it bends upon

  ͟͞

  entering the prism, just as waves would do). But since the advocates

  of that theory had argued (incorrectly) that the spectral splitting was

  due to corruption, Newton’s demonstration that this was not the

  case struck a significant blow in favor of his particle model.

  Newton went on to discover many other facets of light that we

  use today in our understanding of the wave nature of light. He

  showed that every color of light has a unique bend angle when

  passing through a glass prism. He also showed that all objects appear

  to be the same color as the color of the light beam illuminating

  them. And he showed that colored light will not change its color no

  matter how many times it is reflected by or passes through a prism.

  All of these results, including his original result, can be explained

  simply if white light is indeed composed of a collection of different

  colors—that much he got right. But they can’t be explained if light is

  made of different-colored particles. Rather, white light is composed

  of waves of many different wavelengths.

  Newton’s opponents did not give up easily, even in the face of

  Newton’s rising popularity and the death of his chief opponent,

  Hooke. They did not give up even after Newton’s election as

  president of the Royal Society in 1703, the year before he actually

  published his research on light in his epic Opticks. Indeed, the debate

  on the nature of light continued to rage on for over a century.

  Part of the problem with a wave picture of light was the question

  “What is it that light is a wave of, exactly?” And if it is a wave, then

  since all known waves require some medium, what medium does it

  travel in? These questions were sufficiently perplexing that

  practitioners of the wave theory had to resurrect a new invisible

  substance permeating all space, the ether.

  The resolution of this conundrum came, as such resolutions often

  do, from a totally unexpected corner of the physical world, one full

  of sparks, and spinning wheels.

  ͟͟

  When I was a young professor at Yale—in the ancient but huge

  office I was lucky enough to commandeer when an equally ancient

  colleague retired—there was left hanging for me a copy of a

  photograph of Michael Faraday taken in 1861. I have treasured it

  ever since.

  I don’t believe in hero worship, but if I did, Faraday would be up

  there with the best. Perhaps more than any other scientist of the

  nineteenth century, he is responsible for the technology that powers

  our current civilization. Yet he had little formal education and at age

  fourteen became a bookbinder’s apprentice. Later in his career, after

  achieving world recognition for his scientific contributions, he

  insisted on keeping to his humble roots, turning down a knighthood

  and twice turning down the presidency of the Royal Society. Later on

  he refused to advise the British government on the production of

  chemical weapons for use in the Crimean War, citing ethical

  reasons. And for more than thirty-three years he gave a series of

  Christmas lectures at the Royal Institution to excite young people

  about science. What’s not to like?

  Much as one might admire the man, it is the scientist who

  matters here for our story. Faraday’s first scientific lesson is one I tell

  my students: always suck up to your professors. At the age of twenty,

  after completing seven years of apprenticeship as a bookbinder,

  Faraday attended the lectures of the famous chemist Humphry Davy,

  then the head of the Royal Institution. Afterward Faraday presented

  Davy with a three-hundred-page, beautifully bound book containing

  the notes Faraday had taken during the lectures. Within a year,

  Faraday was appointed Davy’s secretary and shortly thereafter got an

  appointment as chemical assistant in the Royal Institution. Later on,

  Faraday learned the same lesson but with the opposite result.

  Following his excitement over some early, quite significant

  experiments that he performed, Faraday accidentally forgot to

  ͟͠

  acknowledge work with Davy in his published results. This

  accidental snub probably resulted in his being reassigned to other

  activities by Davy and delaying his world-changing research by

  several years.

  When reassigned, Faraday had been working on the “hot” area of

  scientific research, the newly discovered connections between

  electricity and magnetism, driven by results of the Danish physicist

  Hans Christian Oersted. These two forces seem quite different, yet

  have odd simi
larities. Electric charges can attract or repel. So can

  magnets. Yet magnets always seem to have two poles, north and

  south, which cannot be isolated, while electric charges can

  individually be positive or negative.

  For some time, scientists and natural philosophers had wondered

  if the two forces might have some hidden connection, and the first

  empirical clue came to Oersted by accident. In 1820, while

  delivering a lecture, Oersted saw that a compass needle was

  deflected when an electric current from a battery was switched on. A

  few months later he followed up on this observation and discovered

  that a current of moving electric charges, which we now commonly

  call an electric current, produced a magnetic attraction that caused

  compass needles to point in a circle around the wire.

  He had blazed a new trail. Word spread quickly among scientists,

  through the Continent and across the English Channel. Moving

  electric charges produced a magnetic force. Could there be other

  connections? Could magnets in turn influence electric charges?

  Scientists searched for such a possibility, without success. Davy

  and another colleague tried to build an electric motor based on the

  connection discovered by Oersted, but failed. Faraday ultimately got

  a wire with a current in it to move around a magnet, which did form

  a crude sort of motor. It was this exciting development that he

  reported without citing Davy’s name.

  ͟͡

  Partly this was mere gamesmanship. No new fundamental

  phenomenon was being uncovered. Perhaps this was the rationale

  for one of my favorite (likely apocryphal) stories about Faraday. It is

  said that William Gladstone, later to be British prime minister, heard

  of Faraday’s laboratory, full of weird devices, and asked in 1850 what

  the practical value of all this study into electricity was. Faraday was

  purported to have replied, “Why, sir, there is every probability that

  you will soon be able to tax it.”

  Apocryphal or not, both great irony and truth are in that witty

  comeback. Curiosity-driven research may seem self-indulgent and

  far from the immediate public good. However, essentially all of our

  current quality of life, for people living in the first world, has arisen

  from the fruits of such research, including all the electric power that

 

‹ Prev