The decline effect and the scientific method : the new yorker
The Decline Effect and the Scientific Method : The New Yorker
THE TRUTH WEARS OFF Is there something wrong with the scientific method?
Many results that are rigorously proved and accepted start shrinking in later studies.
On September 18, 2007, a few dozen neuroscientists, psychiatrists, and drug-company executives gathered in a
hotel conference room in Brussels to hear some startling news. It had to do with a class of drugs known as
atypical or second-generation antipsychotics, which came on the market in the early nineties. The drugs, sold underbrand names such as Abilify, Seroquel, and Zyprexa, had been tested on schizophrenics in several large clinical trials,all of which had demonstrated a dramatic decrease in the subjects’ psychiatric symptoms. As a result, second-generation antipsychotics had become one of the fastest-growing and most profitable pharmaceutical classes. By 2001,Eli Lilly’s Zyprexa was generating more revenue than Prozac. It remains the company’s top-selling drug.
But the data presented at the Brussels meeting made it clear that something strange was happening: the therapeutic
power of the drugs appeared to be steadily waning. A recent study showed an effect that was less than half of thatdocumented in the first trials, in the early nineteen-nineties. Many researchers began to argue that the expensivepharmaceuticals weren’t any better than first-generation antipsychotics, which have been in use since the fifties. “Infact, sometimes they now look even worse,” John Davis, a professor of psychiatry at the University of Illinois atChicago, told me.
Before the effectiveness of a drug can be confirmed, it must be tested and tested again. Different scientists in
different labs need to repeat the protocols and publish their results. The test of replicability, as it’s known, is thefoundation of modern research. Replicability is how the community enforces itself. It’s a safeguard for the creep ofsubjectivity. Most of the time, scientists know what results they want, and that can influence the results they get. Thepremise of replicability is that the scientific community can correct for these flaws.
But now all sorts of well-established, multiply confirmed findings have started to look increasingly uncertain. It’s
as if our facts were losing their truth: claims that have been enshrined in textbooks are suddenly unprovable. Thisphenomenon doesn’t yet have an official name, but it’s occurring across a wide range of fields, from psychology to
http://www.newyorker.com/reporting/2010/12/13/101213fa_fact_lehrer?printable=true¤tPage=all
The Decline Effect and the Scientific Method : The New Yorker
ecology. In the field of medicine, the phenomenon seems extremely widespread, affecting not only antipsychotics butalso therapies ranging from cardiac stents to Vitamin E and antidepressants: Davis has a forthcoming analysisdemonstrating that the efficacy of antidepressants has gone down as much as threefold in recent decades.
For many scientists, the effect is especially troubling because of what it exposes about the scientific process. If
replication is what separates the rigor of science from the squishiness of pseudoscience, where do we put all theserigorously validated findings that can no longer be proved? Which results should we believe? Francis Bacon, theearly-modern philosopher and pioneer of the scientific method, once declared that experiments were essential, becausethey allowed us to “put nature to the question.” But it appears that nature often gives us different answers. Jonathan Schooler was a young graduate student at the University of Washington in the nineteen-eighties when he
discovered a surprising new fact about language and memory. At the time, it was widely believed that the act of
describing our memories improved them. But, in a series of clever experiments, Schooler demonstrated that subjectsshown a face and asked to describe it were much less likely to recognize the face when shown it later than those whohad simply looked at it. Schooler called the phenomenon “verbal overshadowing.”
The study turned him into an academic star. Since its initial publication, in 1990, it has been cited more than four
hundred times. Before long, Schooler had extended the model to a variety of other tasks, such as remembering thetaste of a wine, identifying the best strawberry jam, and solving difficult creative puzzles. In each instance, askingpeople to put their perceptions into words led to dramatic decreases in performance.
But while Schooler was publishing these results in highly reputable journals, a secret worry gnawed at him: it was
proving difficult to replicate his earlier findings. “I’d often still see an effect, but the effect just wouldn’t be asstrong,” he told me. “It was as if verbal overshadowing, my big new idea, was getting weaker.” At first, he assumedthat he’d made an error in experimental design or a statistical miscalculation. But he couldn’t find anything wrongwith his research. He then concluded that his initial batch of research subjects must have been unusually susceptible toverbal overshadowing. (John Davis, similarly, has speculated that part of the drop-off in the effectiveness ofantipsychotics can be attributed to using subjects who suffer from milder forms of psychosis which are less likely toshow dramatic improvement.) “It wasn’t a very satisfying explanation,” Schooler says. “One of my mentors told methat my real mistake was trying to replicate my work. He told me doing that was just setting myself up fordisappointment.”
Schooler tried to put the problem out of his mind; his colleagues assured him that such things happened all the
time. Over the next few years, he found new research questions, got married and had kids. But his replication problemkept on getting worse. His first attempt at replicating the 1990 study, in 1995, resulted in an effect that was thirty percent smaller. The next year, the size of the effect shrank another thirty per cent. When other labs repeated Schooler’sexperiments, they got a similar spread of data, with a distinct downward trend. “This was profoundly frustrating,” hesays. “It was as if nature gave me this great result and then tried to take it back.” In private, Schooler began referringto the problem as “cosmic habituation,” by analogy to the decrease in response that occurs when individuals habituateto particular stimuli. “Habituation is why you don’t notice the stuff that’s always there,” Schooler says. “It’s aninevitable process of adjustment, a ratcheting down of excitement. I started joking that it was like the cosmos washabituating to my ideas. I took it very personally.”
Schooler is now a tenured professor at the University of California at Santa Barbara. He has curly black hair, pale-
green eyes, and the relaxed demeanor of someone who lives five minutes away from his favorite beach. When hespeaks, he tends to get distracted by his own digressions. He might begin with a point about memory, which remindshim of a favorite William James quote, which inspires a long soliloquy on the importance of introspection. Beforelong, we’re looking at pictures from Burning Man on his iPhone, which leads us back to the fragile nature of memory.
Although verbal overshadowing remains a widely accepted theory—it’s often invoked in the context of eyewitness
testimony, for instance—Schooler is still a little peeved at the cosmos. “I know I should just move on already,” hesays. “I really should stop talking about this. But I can’t.” That’s because he is convinced that he has stumbled on aserious problem, one that afflicts many of the most exciting new ideas in psychology.
One of the first demonstrations of this mysterious phenomenon came in the early nineteen-thirties. Joseph Banks
Rhine, a psychologist at Duke, had developed an interest in the possibility of extrasensory perception, or E.S.P. Rhinedevised an experiment featuring Zener cards, a special deck of twenty-five cards printed with one of five differentsymbols: a card was drawn from the deck and the subject was asked to guess the symbol. Most of Rhine’s subjects
http://www.newyorker.com/reporting/2010/12/13/101213fa_fact_lehrer?printable=true¤tPage=all
The Decline Effect and the Scientific Method : The New Yorker
guessed about twenty per cent of the cards correctly, as you’d expect, but an undergraduate named Adam Linzmayeraveraged nearly fifty per cent during his initial sessions, and pulled off several uncanny streaks, such as guessing ninecards in a row. The odds of this happening by chance are about one in two million. Linzmayer did it three times.
Rhine documented these stunning results in his notebook and prepared several papers for publication. But then,
just as he began to believe in the possibility of extrasensory perception, the student lost his spooky talent. Between1931 and 1933, Linzmayer guessed at the identity of another several thousand cards, but his success rate was nowbarely above chance. Rhine was forced to conclude that the student’s “extra-sensory perception ability has gonethrough a marked decline.” And Linzmayer wasn’t the only subject to experience such a drop-off: in nearly every casein which Rhine and others documented E.S.P. the effect dramatically diminished over time. Rhine called this trend the“decline effect.”
Schooler was fascinated by Rhine’s experimental struggles. Here was a scientist who had repeatedly documented
the decline of his data; he seemed to have a talent for finding results that fell apart. In 2004, Schooler embarked on anironic imitation of Rhine’s research: he tried to replicate this failure to replicate. In homage to Rhine’s interests, hedecided to test for a parapsychological phenomenon known as precognition. The experiment itself was straightforward:he flashed a set of images to a subject and asked him or her to identify each one. Most of the time, the response wasnegative—the images were displayed too quickly to register. Then Schooler randomly selected half of the images to beshown again. What he wanted to know was whether the images that got a second showing were more likely to havebeen identified the first time around. Could subsequent exposure have somehow influenced the initial results? Couldthe effect become the cause?
The craziness of the hypothesis was the point: Schooler knows that precognition lacks a scientific explanation. But
he wasn’t testing extrasensory powers; he was testing the decline effect. “At first, the data looked amazing, just aswe’d expected,” Schooler says. “I couldn’t believe the amount of precognition we were finding. But then, as we kepton running subjects, the effect size”—a standard statistical measure—“kept on getting smaller and smaller.” Thescientists eventually tested more than two thousand undergraduates. “In the end, our results looked just like Rhine’s,”Schooler said. “We found this strong paranormal effect, but it disappeared on us.”
The most likely explanation for the decline is an obvious one: regression to the mean. As the experiment is
repeated, that is, an early statistical fluke gets cancelled out. The extrasensory powers of Schooler’s subjects didn’tdecline—they were simply an illusion that vanished over time. And yet Schooler has noticed that many of the datasets that end up declining seem statistically solid—that is, they contain enough data that any regression to the meanshouldn’t be dramatic. “These are the results that pass all the tests,” he says. “The odds of them being random aretypically quite remote, like one in a million. This means that the decline effect should almost never happen. But ithappens all the time! Hell, it’s happened to me multiple times.” And this is why Schooler believes that the declineeffect deserves more attention: its ubiquity seems to violate the laws of statistics. “Whenever I start talking about this,scientists get very nervous,” he says. “But I still want to know what happened to my results. Like most scientists, Iassumed that it would get easier to document my effect over time. I’d get better at doing the experiments, at zeroing inon the conditions that produce verbal overshadowing. So why did the opposite happen? I’m convinced that we can usethe tools of science to figure this out. First, though, we have to admit that we’ve got a problem.”In 1991, the Danish zoologist Anders Møller, at Uppsala University, in Sweden, made a remarkable discovery about
sex, barn swallows, and symmetry. It had long been known that the asymmetrical appearance of a creature was
directly linked to the amount of mutation in its genome, so that more mutations led to more “fluctuating asymmetry.”(An easy way to measure asymmetry in humans is to compare the length of the fingers on each hand.) What Møllerdiscovered is that female barn swallows were far more likely to mate with male birds that had long, symmetricalfeathers. This suggested that the picky females were using symmetry as a proxy for the quality of male genes. Møller’s paper, which was published in Nature, set off a frenzy of research. Here was an easily measured, widelyapplicable indicator of genetic quality, and females could be shown to gravitate toward it. Aesthetics was really aboutgenetics.
In the three years following, there were ten independent tests of the role of fluctuating asymmetry in sexual
selection, and nine of them found a relationship between symmetry and male reproductive success. It didn’t matter ifscientists were looking at the hairs on fruit flies or replicating the swallow studies—females seemed to prefer maleswith mirrored halves. Before long, the theory was applied to humans. Researchers found, for instance, that women
http://www.newyorker.com/reporting/2010/12/13/101213fa_fact_lehrer?printable=true¤tPage=all
The Decline Effect and the Scientific Method : The New Yorker
preferred the smell of symmetrical men, but only during the fertile phase of the menstrual cycle. Other studies claimedthat females had more orgasms when their partners were symmetrical, while a paper by anthropologists at Rutgersanalyzed forty Jamaican dance routines and discovered that symmetrical men were consistently rated as better dancers.
Then the theory started to fall apart. In 1994, there were fourteen published tests of symmetry and sexual selection,
and only eight found a correlation. In 1995, there were eight papers on the subject, and only four got a positive result. By 1998, when there were twelve additional investigations of fluctuating asymmetry, only a third of them confirmedthe theory. Worse still, even the studies that yielded some positive result showed a steadily declining effect size. Between 1992 and 1997, the average effect size shrank by eighty per cent.
And it’s not just fluctuating asymmetry. In 2001, Michael Jennions, a biologist at the Australian National
University, set out to analyze “temporal trends” across a wide range of subjects in ecology and evolutionary biology. He looked at hundreds of papers and forty-four meta-analyses (that is, statistical syntheses of related studies), anddiscovered a consistent decline effect over time, as many of the theories seemed to fade into irrelevance. In fact, evenwhen numerous variables were controlled for—Jennions knew, for instance, that the same author might publish severalcritical papers, which could distort his analysis—there was still a significant decrease in the validity of the hypothesis,often within a year of publication. Jennions admits that his findings are troubling, but expresses a reluctance to talkabout them publicly. “This is a very sensitive issue for scientists,” he says. “You know, we’re supposed to be dealingwith hard facts, the stuff that’s supposed to stand the test of time. But when you see these trends you become a littlemore skeptical of things.”
What happened? Leigh Simmons, a biologist at the University of Western Australia, suggested one explanation
when he told me about his initial enthusiasm for the theory: “I was really excited by fluctuating asymmetry. The earlystudies made the effect look very robust.” He decided to conduct a few experiments of his own, investigatingsymmetry in male horned beetles. “Unfortunately, I couldn’t find the effect,” he said. “But the worst part was thatwhen I submitted these null results I had difficulty getting them published. The journals only wanted confirming data. It was too exciting an idea to disprove, at least back then.” For Simmons, the steep rise and slow fall of fluctuatingasymmetry is a clear example of a scientific paradigm, one of those intellectual fads that both guide and constrainresearch: after a new paradigm is proposed, the peer-review process is tilted toward positive results. But then, after afew years, the academic incentives shift—the paradigm has become entrenched—so that the most notable results arenow those that disprove the theory.
Jennions, similarly, argues that the decline effect is largely a product of publication bias, or the tendency of
scientists and scientific journals to prefer positive data over null results, which is what happens when no effect isfound. The bias was first identified by the statistician Theodore Sterling, in 1959, after he noticed that ninety-seven percent of all published psychological studies with statistically significant data found the effect they were looking for. A“significant” result is defined as any data point that would be produced by chance less than five per cent of the time. This ubiquitous test was invented in 1922 by the English mathematician Ronald Fisher, who picked five per cent asthe boundary line, somewhat arbitrarily, because it made pencil and slide-rule calculations easier. Sterling saw that ifninety-seven per cent of psychology studies were proving their hypotheses, either psychologists were extraordinarilylucky or they published only the outcomes of successful experiments. In recent years, publication bias has mostly beenseen as a problem for clinical trials, since pharmaceutical companies are less interested in publishing results that aren’tfavorable. But it’s becoming increasingly clear that publication bias also produces major distortions in fields withoutlarge corporate incentives, such as psychology and ecology. While publication bias almost certainly plays a role in the decline effect, it remains an incomplete explanation.
For one thing, it fails to account for the initial prevalence of positive results among studies that never even get
submitted to journals. It also fails to explain the experience of people like Schooler, who have been unable to replicatetheir initial data despite their best efforts. Richard Palmer, a biologist at the University of Alberta, who has studiedthe problems surrounding fluctuating asymmetry, suspects that an equally significant issue is the selective reporting ofresults—the data that scientists choose to document in the first place. Palmer’s most convincing evidence relies on astatistical tool known as a funnel graph. When a large number of studies have been done on a single subject, the datashould follow a pattern: studies with a large sample size should all cluster around a common value—the true result—whereas those with a smaller sample size should exhibit a random scattering, since they’re subject to greater samplingerror. This pattern gives the graph its name, since the distribution resembles a funnel.
http://www.newyorker.com/reporting/2010/12/13/101213fa_fact_lehrer?printable=true¤tPage=all
The Decline Effect and the Scientific Method : The New Yorker
The funnel graph visually captures the distortions of selective reporting. For instance, after Palmer plotted every
study of fluctuating asymmetry, he noticed that the distribution of results with smaller sample sizes wasn’t random atall but instead skewed heavily toward positive results. Palmer has since documented a similar problem in several othercontested subject areas. “Once I realized that selective reporting is everywhere in science, I got quite depressed,”Palmer told me. “As a researcher, you’re always aware that there might be some nonrandom patterns, but I had no ideahow widespread it is.” In a recent review article, Palmer summarized the impact of selective reporting on his field:“We cannot escape the troubling conclusion that some—perhaps many—cherished generalities are at best exaggeratedin their biological significance and at worst a collective illusion nurtured by strong a-priori beliefs often repeated.”
Palmer emphasizes that selective reporting is not the same as scientific fraud. Rather, the problem seems to be one
of subtle omissions and unconscious misperceptions, as researchers struggle to make sense of their results. StephenJay Gould referred to this as the “shoehorning” process. “A lot of scientific measurement is really hard,” Simmons toldme. “If you’re talking about fluctuating asymmetry, then it’s a matter of minuscule differences between the right andleft sides of an animal. It’s millimetres of a tail feather. And so maybe a researcher knows that he’s measuring a goodmale”—an animal that has successfully mated—“and he knows that it’s supposed to be symmetrical. Well, that act ofmeasurement is going to be vulnerable to all sorts of perception biases. That’s not a cynical statement. That’s just theway human beings work.”
One of the classic examples of selective reporting concerns the testing of acupuncture in different countries. While
acupuncture is widely accepted as a medical treatment in various Asian countries, its use is much more contested inthe West. These cultural differences have profoundly influenced the results of clinical trials. Between 1966 and 1995,there were forty-seven studies of acupuncture in China, Taiwan, and Japan, and every single trial concluded thatacupuncture was an effective treatment. During the same period, there were ninety-four clinical trials of acupuncture inthe United States, Sweden, and the U.K., and only fifty-six per cent of these studies found any therapeutic benefits. AsPalmer notes, this wide discrepancy suggests that scientists find ways to confirm their preferred hypothesis,disregarding what they don’t want to see. Our beliefs are a form of blindness.
John Ioannidis, an epidemiologist at Stanford University, argues that such distortions are a serious issue in
biomedical research. “These exaggerations are why the decline has become so common,” he says. “It’d be really greatif the initial studies gave us an accurate summary of things. But they don’t. And so what happens is we waste a lot ofmoney treating millions of patients and doing lots of follow-up studies on other themes based on results that aremisleading.” In 2005, Ioannidis published an article in the Journal of the American Medical Association that looked atthe forty-nine most cited clinical-research studies in three major medical journals. Forty-five of these studies reportedpositive results, suggesting that the intervention being tested was effective. Because most of these studies wererandomized controlled trials—the “gold standard” of medical evidence—they tended to have a significant impact onclinical practice, and led to the spread of treatments such as hormone replacement therapy for menopausal women anddaily low-dose aspirin to prevent heart attacks and strokes. Nevertheless, the data Ioannidis found were disturbing: ofthe thirty-four claims that had been subject to replication, forty-one per cent had either been directly contradicted orhad their effect sizes significantly downgraded.
The situation is even worse when a subject is fashionable. In recent years, for instance, there have been hundreds
of studies on the various genes that control the differences in disease risk between men and women. These findingshave included everything from the mutations responsible for the increased risk of schizophrenia to the genesunderlying hypertension. Ioannidis and his colleagues looked at four hundred and thirty-two of these claims. Theyquickly discovered that the vast majority had serious flaws. But the most troubling fact emerged when he looked atthe test of replication: out of four hundred and thirty-two claims, only a single one was consistently replicable. “Thisdoesn’t mean that none of these claims will turn out to be true,” he says. “But, given that most of them were donebadly, I wouldn’t hold my breath.”
According to Ioannidis, the main problem is that too many researchers engage in what he calls “significance
chasing,” or finding ways to interpret the data so that it passes the statistical test of significance—the ninety-five-per-cent boundary invented by Ronald Fisher. “The scientists are so eager to pass this magical test that they start playingaround with the numbers, trying to find anything that seems worthy,” Ioannidis says. In recent years, Ioannidis hasbecome increasingly blunt about the pervasiveness of the problem. One of his most cited papers has a deliberatelyprovocative title: “Why Most Published Research Findings Are False.”
The problem of selective reporting is rooted in a fundamental cognitive flaw, which is that we like proving
http://www.newyorker.com/reporting/2010/12/13/101213fa_fact_lehrer?printable=true¤tPage=all
The Decline Effect and the Scientific Method : The New Yorker
ourselves right and hate being wrong. “It feels good to validate a hypothesis,” Ioannidis said. “It feels even betterwhen you’ve got a financial interest in the idea or your career depends upon it. And that’s why, even after a claim hasbeen systematically disproven”—he cites, for instance, the early work on hormone replacement therapy, or claimsinvolving various vitamins—“you still see some stubborn researchers citing the first few studies that show a strongeffect. They really want to believe that it’s true.”
That’s why Schooler argues that scientists need to become more rigorous about data collection before they publish.
“We’re wasting too much time chasing after bad studies and underpowered experiments,” he says. The current“obsession” with replicability distracts from the real problem, which is faulty design. He notes that nobody even triesto replicate most science papers—there are simply too many. (According to Nature, a third of all studies never evenget cited, let alone repeated.) “I’ve learned the hard way to be exceedingly careful,” Schooler says. “Every researchershould have to spell out, in advance, how many subjects they’re going to use, and what exactly they’re testing, andwhat constitutes a sufficient level of proof. We have the tools to be much more transparent about our experiments.”
In a forthcoming paper, Schooler recommends the establishment of an open-source database, in which researchers
are required to outline their planned investigations and document all their results. “I think this would provide a hugeincrease in access to scientific work and give us a much better way to judge the quality of an experiment,” Schoolersays. “It would help us finally deal with all these issues that the decline effect is exposing.”Although such reforms would mitigate the dangers of publication bias and selective reporting, they still wouldn’t
erase the decline effect. This is largely because scientific research will always be shadowed by a force that can’t
be curbed, only contained: sheer randomness. Although little research has been done on the experimental dangers ofchance and happenstance, the research that exists isn’t encouraging.
In the late nineteen-nineties, John Crabbe, a neuroscientist at the Oregon Health and Science University, conducted
an experiment that showed how unknowable chance events can skew tests of replicability. He performed a series ofexperiments on mouse behavior in three different science labs: in Albany, New York; Edmonton, Alberta; andPortland, Oregon. Before he conducted the experiments, he tried to standardize every variable he could think of. Thesame strains of mice were used in each lab, shipped on the same day from the same supplier. The animals were raisedin the same kind of enclosure, with the same brand of sawdust bedding. They had been exposed to the same amount ofincandescent light, were living with the same number of littermates, and were fed the exact same type of chow pellets. When the mice were handled, it was with the same kind of surgical glove, and when they were tested it was on thesame equipment, at the same time in the morning.
The premise of this test of replicability, of course, is that each of the labs should have generated the same pattern
of results. “If any set of experiments should have passed the test, it should have been ours,” Crabbe says. “But that’snot the way it turned out.” In one experiment, Crabbe injected a particular strain of mouse with cocaine. In Portlandthe mice given the drug moved, on average, six hundred centimetres more than they normally did; in Albany theymoved seven hundred and one additional centimetres. But in the Edmonton lab they moved more than five thousandadditional centimetres. Similar deviations were observed in a test of anxiety. Furthermore, these inconsistencies didn’tfollow any detectable pattern. In Portland one strain of mouse proved most anxious, while in Albany another strainwon that distinction.
The disturbing implication of the Crabbe study is that a lot of extraordinary scientific data are nothing but noise.
The hyperactivity of those coked-up Edmonton mice wasn’t an interesting new fact—it was a meaningless outlier, aby-product of invisible variables we don’t understand. The problem, of course, is that such dramatic findings are alsothe most likely to get published in prestigious journals, since the data are both statistically significant and entirelyunexpected. Grants get written, follow-up studies are conducted. The end result is a scientific accident that can takeyears to unravel.
This suggests that the decline effect is actually a decline of illusion. While Karl Popper imagined falsification
occurring with a single, definitive experiment—Galileo refuted Aristotelian mechanics in an afternoon—the processturns out to be much messier than that. Many scientific theories continue to be considered true even after failingnumerous experimental tests. Verbal overshadowing might exhibit the decline effect, but it remains extensively reliedupon within the field. The same holds for any number of phenomena, from the disappearing benefits of second-generation antipsychotics to the weak coupling ratio exhibited by decaying neutrons, which appears to have fallen bymore than ten standard deviations between 1969 and 2001. Even the law of gravity hasn’t always been perfect at
http://www.newyorker.com/reporting/2010/12/13/101213fa_fact_lehrer?printable=true¤tPage=all
The Decline Effect and the Scientific Method : The New Yorker
predicting real-world phenomena. (In one test, physicists measuring gravity by means of deep boreholes in the Nevadadesert found a two-and-a-half-per-cent discrepancy between the theoretical predictions and the actual data.) Despitethese findings, second-generation antipsychotics are still widely prescribed, and our model of the neutron hasn’tchanged. The law of gravity remains the same.
Such anomalies demonstrate the slipperiness of empiricism. Although many scientific ideas generate conflicting
results and suffer from falling effect sizes, they continue to get cited in the textbooks and drive standard medicalpractice. Why? Because these ideas seem true. Because they make sense. Because we can’t bear to let them go. Andthis is why the decline effect is so troubling. Not because it reveals the human fallibility of science, in which data aretweaked and beliefs shape perceptions. (Such shortcomings aren’t surprising, at least for scientists.) And not because itreveals that many of our most exciting theories are fleeting fads and will soon be rejected. (That idea has been aroundsince Thomas Kuhn.) The decline effect is troubling because it reminds us how difficult it is to prove anything. Welike to pretend that our experiments define the truth for us. But that’s often not the case. Just because an idea is truedoesn’t mean it can be proved. And just because an idea can be proved doesn’t mean it’s true. When the experimentsare done, we still have to choose what to believe. ♦
To get more of The New Yorker's signature mix of politics, culture and the arts:
http://www.newyorker.com/reporting/2010/12/13/101213fa_fact_lehrer?printable=true¤tPage=all
PERBEDAAN KADAR MATRIX METALLOPROTEINASE-8SETELAH SCALING DAN PEMBERIAN TETRASIKLIN PADA(The difference of Matrix metalloproteinase-8 levels after scaling andtetracycline addition of chronic periodontitis)Melok Aris WahyukundariDepartemen PeriodonsiaFakultas Kedokteran Gigi Universitas AirlanggaSurabaya – IndonesiaMatrix metalloproteinase, especially MMP-8 is the key indicator related to pe
Chronicle of a Pandemic Foretold: Lessons from the 2009 Influenza Epidemic Available at SSRN: http://ssrn.com/abstract=1398445. Executive Summary The A(H1N1) influenza epidemic provided the first indication of the effectiveness of the pandemic preparations that countries and international organizations initiated in the wake of the 2003 SARS epidemic. In the case of SARS, China was