[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]
Fw: Epidemiology Epedemic
----- Original Message -----
From: jjcohen
To: Jerry Cohen
Sent: Friday, July 02, 2004 4:38 PM
Subject: Fw: Epidemiology Epedemic
----- Original Message -----
From: jjcohen
To: radsafe@list.vanderbilt.edu
Sent: Friday, July 02, 2004 4:08 PM
Subject: Fw: Epidemiology Epedemic
In previous discussions on radsafe, concerns were expressed regarding how
epidemiology had been applied in determining effects of exposure to radon,
radioactivity , and 'hazardous" materials in general.
It appeared, at least to me, that certain epidemiological determinations
were being made that defied common sense. Recently, on another website, I
came across the following review and thought the group might be interest.
Anyone have any thoughts on the subject?? Jerry Cohen
----------------------------------------------------------------------------
-----------------------
An Epidemic of Epidemiology
by Rob Lyons
Fifty years ago, we discovered that smoking is bad for us. In 1954, Austin
Bradford Hill and Richard Doll published a preliminary report on a study
showing the very strong correlation between smoking and premature mortality
(1).
However, this classic study has in many ways sent medical science up a blind
alley. While the dangers of smoking have been demonstrated in numerous
subsequent studies, the attempts to find the New Smoking - another example
of an environmental or lifestyle factor that causes substantial health
problems - have largely failed. But the many pieces of junk science that
have been produced in the process have provided the ammunition for
unwarranted health scares too numerous to mention.
This state of affairs is well described in John Brignell's new book The
Epidemiologists. Hill and Doll were given the task of trying to find out why
cases of lung cancer had increased 15-fold in only 25 years. Their first
attempt was to ask 649 lung cancer patients, and 649 matched controls, about
their habits. What they found was a correlation between smoking and lung
cancer, albeit not a very strong one. However, it was strong enough to
warrant a fuller study, starting with a large group of healthy individuals,
assessing their smoking habits and then monitoring them to see what diseases
they developed.
This study began in 1951. Their method was to write to every doctor in the
country - around 35,000 doctors replied, of whom only 17 per cent were
lifelong non-smokers (how times change). The doctors were asked just a few
questions about their smoking habits. Three years later, Hill and Doll
published their first analysis of the results, and were already able to
indicate how strong the link was between smoking and lung cancer.
What they found was that persistent smokers were 24 times more likely to
develop lung cancer than non-smokers. Moreover, the risk of death from heart
disease in any particular year was roughly doubled. This study has been
followed up every few years, and these results have been confirmed time and
again.
Hill made it clear, however, that such a study had to comply with some
pretty strict criteria in order to be considered valid. These criteria are
worth restating, because they stand in sharp contrast to the bulk of
epidemiological research:
1. Strength
Is the association strong enough that we can rule out other factors?
2. Consistency
Have the results been replicated by different researchers, and under
different conditions?
3. Specificity
Is the exposure associated with a very specific disease as opposed to a wide
range of diseases?
4. Temporality
Did the exposure proceed the disease?
5. Biological gradient
Are increasing exposures associated with increasing risk of disease?
6. Plausibility
Is there a credible scientific mechanism that can explain the association?
7. Coherence
Is the association consistent with the natural history of the disease?
8. Experimental evidence
Does a physical intervention show results consistent with the association?
9. Analogy
Is there a similar result to which we can draw a relationship?
Above all, as Brignell emphasizes, correlation does not prove causation. He
draws an analogy with growing tomatoes and fertilizer. It can easily be
shown that increasing use of fertilizer will increase tomato yields. But
fertilizer does not cause tomatoes, it merely promotes the process of
growth. The same goes for smoking and lung cancer. Smoking may massively
promote the growth of lung cancer, but it does not cause the tumours. Hill
and Doll had nothing to say about why cancer occurs in the first place.
Nonetheless, it is an entirely reasonable conclusion to draw that smokers
will, on average, die younger than non-smokers, and we do not need to know
the precise mechanism to conclude that giving up smoking is prudent from a
health viewpoint.
What is not reasonable is the response to this one, classic study. First, it
has provided the justification for state intervention in lifestyle in a
previously unprecedented way. Secondly, it has encouraged the proliferation
of other studies, which make grand statements about disease based on
correlations far weaker than those found by Hill and Doll.
Brignell's book is a handy demolition of the science and statistics behind
this epidemic of epidemiology. He shows how statistical tests were
originally developed, based on certain assumptions. However, these
assumptions have long since been forgotten, so that meeting certain abstract
criteria has been elevated above whether the results are actually of any
real-world importance.
The most important of these is the test for statistical significance. The
idea behind this is that patterns can be found in any set of random results.
For example, in the spiked office, there are a number of people who were
born in May or June, but none were born in July. It would be possible to
draw the conclusion that there is something special about being born in May
or June that predisposes people to become journalists. This would be a
bizarre conclusion to draw from just a handful of people. In fact, the
spread of birthdays is completely coincidental.
In research it is therefore useful to have a preliminary statistical test of
results, to see how likely it is that they could be due to blind chance. The
usual benchmark is that if the chances of a set of results being
coincidental are less than five per cent, it is reasonable to go on to
assess whether the results are actually meaningful.
First, just because a study passes this test does not mean its results
aren't a complete coincidence. In fact, by definition, five per cent of
studies could pass this test even though the results are meaningless.
Secondly, just because the results are statistically significant doesn't
mean they are practically significant. Brignell gives the example of a book
called The Causes of Cancer, written by Richard Doll and Richard Peto.
Illustrating Doll's fall from previous high standards, the book describes
some deaths of people in their 80s and 90s as 'premature'.
The public health agenda is justified by research that is often completely
worthless
These days, however, it seems that any result that passes this 'p-test' is
increasingly regarded as significant. Five per cent sounds like a low risk
of results being meaningless, until you realise that researchers often
plough through many, potential risk factors (what Brignell calls a 'data
dredge'), look for an apparently significant result, then try to speculate
some kind of mechanism to explain it, no matter how bizarre. So a test
designed as an initial filter to weed out spurious results is used to give
credence to them.
Thus he provides a huge list of different factors that have, at one time or
other, been accused of causing cancer: abortion, acetaldehyde, acrylamide,
acrylontiril, agent orange, alar, alcohol, air pollution, aldrin, alfatoxin,
arsenic, asbestos, asphalt fumes, atrazine, AZT.and that's just the letter
'A'.
There are also a number of techniques in epidemiology for imposing
assumptions on to data. The best of these is trend fitting. No set of data
will exactly fit a pattern but often a clear trend can be found nonetheless.
However, many studies appear to resort to drawing a line through an
apparently unconnected series of measurements to demonstrate an underlying
effect.
Epidemiology can be an effective tool when applied to the spread of
infectious disease. Unfortunately, there really isn't anything like enough
infectious disease in the developed world to justify the existence of so
many departments and researchers. In fact, the overwhelming cause of death
in the developed world is old age - a factor that is, incredibly, frequently
ignored by researchers. A person in their eighties is a thousand times more
likely to develop cancer than someone in their thirties. This factor is so
powerful that for most of the causes of disease studied, a very minor
underestimation of the effect of age can wipe out any putative effect from
the factor in hand.
Age is obvious - but many other confounding factors are not. Therefore, we
return to Hill's first criterion: to be sure there is actually something
going on, the effect must be strong. Otherwise, any apparent effect may
prove to be entirely illusory.
A topical example of this is passive smoking, and in particular what
Brignell calls 'the greatest scientific fraud ever'. In 1992, the US
Environmental Protection Agency published a meta-study, bringing together
many other studies on passive smoking. Unfortunately, the results were
negative. It appeared that passive smoking was not a health risk at all.
Mere facts could not be allowed to get in the way of a health scare, so some
imagination was applied to the problem. One negative study was removed - but
the meta-study still produced no statistically significant result.
So the goalposts were not so much moved as widened. The organisation found
that there was a greater than five per cent chance that the results were
coincidental, but less than 10 per cent - so they accepted them anyway. In
other words, the EPA accepted a bigger risk that the effect they found was
purely due to chance, quite at odds with standard practice.
The increased risk of lung cancer they found - 19 per cent - was frankly too
small to have been conceivably detected given the methods they used. There
are lots of ways in which inaccuracy could have crept into this final
result. For example, is it really possible to merge the results of many
different studies, all with different methodologies and subjects,
accurately? How could someone's actual exposure to environmental smoke be
measured over the course of years? Were all the people who said that they
were non-smokers absolutely honest? As indicated above, were other possible
contributory factors such as age, gender and income controlled for
accurately?
We can be pretty confident about Hill and Doll's conclusions about lung
cancer because the effect they found is massive - an increased risk of 2400
percent. To suggest that such a small effect as 19 per cent could be
accurately measured in this way is like trying to time a race with a
sundial.
That has not prevented smoking being banned in public places on the grounds
that thousands of people might die from inhaling second-hand smoke. The
public health agenda is therefore driven and justified by research that is
more often than not completely worthless.
It is undoubtedly the case that Hill and Doll's study has caused people to
give up smoking and extended many lives as a result. But it has also
inspired a heap of unnecessary panics based on dodgy research, and public
health campaigners only too willing to tell us how to live our lives.
************************************************************************
You are currently subscribed to the Radsafe mailing list. To
unsubscribe, send an e-mail to Majordomo@list.vanderbilt.edu Put the
text "unsubscribe radsafe" (no quote marks) in the body of the e-mail,
with no subject line. You can view the Radsafe archives at
http://www.vanderbilt.edu/radsafe/