The Half-Life of Facts (4 page)

Read The Half-Life of Facts Online

Authors: Samuel Arbesman

BOOK: The Half-Life of Facts
5.53Mb size Format: txt, pdf, ePub

As one generous Nobel laureate in chemistry put it: “It helps a young man to be senior author, first author, and doesn’t detract from the credit that I get if my name is farther down the list.” On the other hand, those peers of Nobel laureates who were not as successful tried to maintain first authorship for themselves far more often, garnering more glory for themselves. By their forties, Nobel laureates are first authors on only 26 percent of their papers, as compared to their less accomplished contemporaries, who are first authors 56 percent of the time. Nicer people are indeed more creative, more successful, and even more likely to win Nobel prizes.

These regular patterns of scientists seem evident enough, at least when we look at whole populations of researchers. But what of regularities related to knowledge itself and how it’s created? To understand this, we need to begin thinking about asteroids.

.   .   .

ARTHUR
C. Clarke was my favorite writer when I was a teenager. His visionary approach to the world around us, and his imaginary futures, were constructed in great detail, and they often pointed to the most positive aspects of humanity. While he is most famous for
2001: A Space Odyssey
, he wrote dozens of books and many more essays and short stories.

In one of his other well-known books,
Rendezvous with Rama
, which was published in 1973, a large cylindrical starship flies through our solar system, and a team of astronauts is sent to unravel its secrets. That starship,
Rama
, was initially thought to be an asteroid; however, it is detected by a series of automated telescopes known as
SPACEGUARD
. These had been put into place after a meteor smashed into northern Italy in 2077, leading humanity to create a sort of early warning system for any objects that might cross Earth’s path and potentially threaten our well-being.

When, a number of years after the book’s publication, a project like this was actually proposed, its creators paid homage to Clarke and named it the Spaceguard Survey. It uses a variety of discovery methods, including automatic detection of objects in space. In addition to being a real-life incarnation of science fiction, it is also the vanguard of a whole new way of doing science: automatically. One portion of this program, known as Spacewatch, uses automated image processing to detect what are termed NEOs, or near earth objects. In 1992, it was responsible for the first automated discovery of a comet, which now goes by the unwieldy name of C/1992 J1.

Automated science is being done in fields from biology to astronomy to theoretical mathematics, all using computers to make new discoveries and identify new facts. We have projects that find new species by looking at the genetic sequences in water scooped out of the oceans, and ones that allow chemists to automatically discover ways to synthesize specific chemicals. There is even a Web site named TheoryMine to which anyone can go and get a novel mathematical theorem created by a sophisticated computer program that automatically generates a mathematical proof and have it named after you or a loved one. This is automated discovery combined with vanity plates.

But that’s not the only big thing going on right now: There is a movement known as citizen science. Everyday individuals are becoming involved in actual scientific discovery. How is this possible?
The principle is rather elegant. While computers are good at lots of things, from adding numbers to counting words in a document, they are often very bad at many simple things: We are still way ahead of computers in labeling photographs or even reading fuzzy or blurred text. This computer limitation, in addition to providing a stumbling block to any robots whose route to global domination relies on caption creation, has created an entirely new field of computer science known as human computation: Simple tasks are given to lots of people to perform, often either for a small amount of money or because someone has cleverly hidden the task in a game. One of the most well-known examples of these are the distorted words we often have to read correctly in order to prove our humanity to a Web site. Rather than simply being an inconvenience, they are now being exploited to actually help digitize such works as the
New York Times
archives. By pairing a distorted known word with one that computers are unable to decipher, everyday users who can read these words are helping bring newspapers and books into digital formats.

Scientists are beginning to use this sort of human computation. These researchers are relying on citizen scientists to help them look through large amounts of data, most of which is too difficult for a computer (or even a single person) to comb through. One example is Galaxy Zoo, in which scientists gave participants pictures of galaxies to help classify them. The participants weren’t experts; they were interested individuals who participated in a minutes-long tutorial and were simply interested in space, or wanted to help scientific progress.

Several intrepid scientists turned a fiendishly difficult problem—how to predict what shapes proteins will fold into based on their chemical makeup—into a game. They found that the best players of a simple online game known as Foldit are actually better than our best computers.

Our pattern-detection abilities, and other quirks of how our brains work, still give us the lead in many tasks that are required for new knowledge. So, in truth, the facts that are changing are not
changing simply without the involvement of the general population. We are part of the scientific process now. Each of us, not just scientists, inventors, or even explorers, are able to be a part of the process of creating knowledge.

We are living during an auspicious time with respect to knowledge: For the first time, not only has massive computational power allowed for much information related to scientific discovery to be digitized—the amount of scientific data available online for analysis is simply staggering—but discoveries are actually occurring automatically through computational discovery, and also in a widely distributed fashion, through citizen science.

These combined forces, in addition to changing how new scientific knowledge is generated, have enabled us to obtain massive amounts of data on the properties of scientific discoveries and how they are found. This has led to what I, along with one of my collaborators, Nicholas Christakis, have taken to calling
eurekometrics
. Eurekometrics is about studying scientific discoveries themselves. More traditional scientometric approaches that use citations are still very important. They can teach us how scientists collaborate, measure the impact of scientific research, and chart how scientific knowledge grows, but they often tell us nothing about the content of the discoveries themselves, or their properties. For example, rather than looking at the properties of the articles appearing in plant biology journals, we can instead look at the properties of the plant species that have been discovered.

One simple example of eurekometrics—and one that I was involved in—is examining how discoveries become more difficult over time.

.   .   .

IF
you look back in history you can get the impression that scientific discoveries used to be easy. Galileo rolled objects down slopes; Robert Hooke played with a spring to learn about elasticity; Isaac Newton poked around his own eye with a darning needle to understand color perception. It took creativity and knowledge (and perhaps a
lack of squeamishness or regard for one’s own well-being) to ask the right questions, but the experiments themselves could be very simple. Today, if you want to make a discovery in physics, it helps to be part of a ten-thousand-member team that runs a multibillion-dollar atom smasher. It takes ever more money, more effort, and more people to find out new things.

Until recently, no one actually tried to measure the increasing difficulty of discovery. It certainly seems to be getting harder, but how much harder? How fast does it change?

I approached this question in a eurekometric frame of mind and looked at three specific areas of science: mammal species, asteroids, and chemical elements. These areas have two primary benefits: In addition to being from different fields of science, many have clear discovery data going back hundreds of years. The first mammals discovered after the creation of the classification system developed by Carl Linnaeus date to the 1760s. The first asteroid discovered, Ceres, was found in 1801 (and was actually large enough to be thought a planet). And the first modern chemical element discovered (I ignored such elements as lead and gold, which have been known since ancient times) was phosphorus, in 1669.

As I began thinking about how to understand how discoveries get harder, I settled on size. I assumed that size is a good proxy for how easy it is to discover something: The smaller a creature or asteroid is, the harder it is to discover; in chemistry, the reverse is true, and the largest elements are the hardest to create and detect, so I used inverse size. Based on this, I plotted the average ease of discovery over time.

What I found, using this simple proxy for difficulty, was a clear pattern of how discoveries occur: Each dataset adhered to a curve with the same basic shape. In every case, the ease of discovery went down, and in every case, it was an exponential decay.

What this means is that the ease of discovery doesn’t drop by the same amount every year—it declines by the same fraction each year, a sort of reverse compound interest. For example, the sizes of asteroids discovered annually get 2.5 percent smaller each year. In the first few years, the ease of discovery drops off quickly; after
early researchers pick the low-hanging fruit, it continues to “decay” for a long time, becoming slightly harder without ever quite becoming impossible.

There is even a similarity in one view of medicine. As Tyler Cowen, an economist at George Mason University, has noted, if you tally the number of major advances, or definitive moments, in modern medicine (as chronicled by James Le Fanu) in each decade of the middle of the twentieth century, you get an eventual decline: “In the 1940s there are six such moments, seven moments in the 1950s, six moments in the 1960s, a moment in 1970 and 1971 each, and from 1973 [to] 1998, a twenty-five-year period, there are only seven moments in total.”

But here’s the wonderful thing: The output of discovery keeps marching on in each of the areas I examined. We keep on discovering more asteroids, new mammals, and increasingly exotic chemical elements, even as each successive discovery becomes harder. These all occur at different rates—we find asteroids much faster than new types of mammal, for example—but we aren’t at the end of discovery in any of these areas. In fact, I only know of one area where scientific research has exhausted all discoveries: the “field” of the discovery of new major internal organs.

The trajectory of discovery in human anatomy began in prehistoric times with the discoveries of hearts and lungs, the organs that are rather hard to miss, especially after seeing your colleague disemboweled by a mastodon. This initial flowering of discovery was followed by that of more subtle organs, such as the pituitary gland. But in 1880, a Swedish medical student named Ivar Sandström discovered the parathyroid gland, and the final major internal organ in humans was discovered. That was it. The field’s age of discovery was over.

But science as a whole proceeds apace. We pour in more effort, more manpower, and greater ingenuity into further discovery and knowledge. A simple example is one of the first quantities to be studied in the field of scientometrics: the number of scientists over time. The first U.S. PhDs were granted by Yale University in 1861. Since that time the number of scientists in the United States and
throughout the world has increased rapidly. For example, the membership of scientists in the Federation of American Societies for Experimental Biology increased from fewer than five hundred in the year 1920 to well over fifty thousand by the late 1990s. This hundredfold increase is extremely rapid in a period of less than eighty years and is indicative of the increase in scientific power through sheer numbers of scientists.

In fact, if you uttered the statement “Eighty percent of all the scientists who have ever lived are alive today” nearly anytime in the past three hundred years, you’d be right. This has allowed more research to be done by larger scientific teams. Not only that, but higher-impact research is done by teams with many more scientists. Of course, growth like this is not sustainable—a long exponential increase in the number of scientists means that at some point the number of scientists would need to exceed the number of humans on Earth. While this may almost be the case on Krypton, Superman’s home planet, where the whole population seems to consist entirely of scientists, I don’t see this happening on Earth anytime soon. But this rapid growth demonstrates that scientific discovery is by no means anywhere near finished.

For example, in pharmaceutical research, drug companies counter the decreasing number of drugs created per dollar spent by pouring more money into drug discovery. As science grows exponentially more difficult in some areas, affordable technology often proceeds along a similar curve: an exponential increase in computer processing power means that problems once considered hard, such as visualizing fractals, proving certain mathematical theorems, or simulating entire populations, can now be done quite easily. Some scientists arrive at new discoveries without a significant investment in resources by becoming more clever and innovative. When Stanley Milgram did his famous “six degrees of separation” experiment, the one that showed that everyone on Earth was much more closely linked than we imagined, he did it by using not much more than postcards and stamps.

When one area of research becomes difficult, the scientists in
that field either rise to the challenge by investing greater effort or shift their focus of inquiry. Physicists have moved into biology, raising new questions that no one had thought to ask before. Mathematicians and computer scientists have turned their formulas and algorithms to the social sciences and unleashed basic new discoveries about the way societies operate. Or scientists figure out ways to make the hard questions much easier, whether by importing techniques from other areas or inventing new methods.

Other books

Get Lucky by Lorie O'clare
Science Fair by Dave Barry, Ridley Pearson
The Madman Theory by Ellery Queen
Shadowed By Wings by Janine Cross
H. M. S. Ulysses by Alistair MacLean
The Future We Left Behind by Mike A. Lancaster
Prometheus Rising by Aaron Johnson