Next Article in Journal
Challenging the “Inoffensiveness” of Regular Cannabis Use by Its Associations with Other Current Risky Substance Use—A Census of 20-Year-Old Swiss Men
Next Article in Special Issue
Advances in Identifying Beryllium Sensitization and Disease
Previous Article in Journal
Publisher’s Note – Received Date Error, Int. J. Environ. Res. Public Health 2009, 6, 3105-3114
Previous Article in Special Issue
Examining the Dynamic Association of BMI and Mortality in the Framingham Heart Study
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Epidemiological Methods: About Time

by
Helena Chmura Kraemer
1,2
1
Department of Psychiatry and Behavioral Sciences, Stanford University, 1116 Forest Avenue, Palo Alto, CA 94301, USA
2
Department of Psychiatry, University of Pittsburgh, 3811 O'Hara Street, Pittsburgh, PA 15213, USA
Int. J. Environ. Res. Public Health 2010, 7(1), 29-45; https://doi.org/10.3390/ijerph7010029
Submission received: 29 October 2009 / Accepted: 24 December 2009 / Published: 31 December 2009
(This article belongs to the Special Issue Advances in Epidemiology)

Abstract

:
Epidemiological studies often produce false positive results due to use of statistical approaches that either ignore or distort time. The three time-related issues of focus in this discussion are: (1) cross-sectional vs. cohort studies, (2) statistical significance vs. public health significance, and (3), how risk factors “work together” to impact public health significance. The issue of time should be central to all thinking in epidemiology research, affecting sampling, measurement, design, analysis and, perhaps most important, the interpretation of results that might influence clinical and public-health decision-making and subsequent clinical research.

1. Introduction

Epidemiology has been defined as the “study of the distribution and determinants of health-related states or events in specified populations, and the application of this study to control of health problems” [1] (Page 55). “Distribution” refers to incidence or prevalence of disorders in specified time periods. “Determinants” refers to risk factors (some causal, some not) for disorders, factors that can be shown to identify individuals who at a later time are more likely to have the disorder. “Application” refers in part to prevention of disorders in the subsequent period of time by manipulating causal risk factors among those as yet without the disorder. All such tasks are crucially dependent on considerations of time. While other types of questions also fall within the bailiwick of epidemiology research, the unique contribution of epidemiology research is guidance based on observational studies provided for the prediction and prevention of disorders, both crucially time-related. Yet such epidemiological studies have often produced false positive results [25], even studies that seem well and carefully designed, executed, and analyzed. At least part of that problem relates to the statistical methods used in epidemiology, many of which ignore or distort time.
In what follows, we address three interlocking areas of such concern in observational studies in human populations: (1) cross-sectional vs. prospective longitudinal (cohort) studies and their design; (2) statistical significance vs. public health significance and (3), how risk factors “work together” to impact public health significance. None of these issues is completely novel, but the problems they generate continue to affect epidemiological studies, and it is worth considering why this is so.

2. Cross-Sectional and Prospective Longitudinal (Cohort) Studies

To show that some factor is a “risk factor” for a disorder, it must be shown both that: (a) the factor precedes onset of the disorder, and (b) it is correlated with the disorder [6]. A factor that is correlated with presence/absence of a disorder in a cross-sectional study may be a symptom of, or a result of, the disorder, not a risk factor for it. Presence or absence of the disorder may be indicated either by prevalence or incidence within a specified time period.
For a disorder that is chronic and persistent after onset, prevalence at a certain age equals the incidence between birth and that age for those that survive to that age. Otherwise, for a disorder from which there may be remissions or recoveries, or one associated with removals from the population, cross sectional correlations may relate as much to the association of factors related to treatment availability and response, or inconsistency of expression over time, as to risk of incurring the disorder. Consequently cross-sectional studies investigating correlates of prevalence of episodic disorders are of limited use in identifying determinants of those disorders or ways of preventing them, although they may be vital in setting the stage for prospective studies to accomplish those purposes.
One exception relates to the detection of “fixed markers” [6], i.e., risk factors that are fixed over the course of the lifetime of individuals, e.g., gender, ethnicity/race, year of birth, and, in general, genotype. Since fixed markers exist at birth and remain unchanged over the lifetime of the individual, these are temporally precedent to onset of all disorders during the individual’s lifetime. Thus to show correlation between a fixed marker and presence/absence of a disorder in a cross-sectional study is to establish that the variable is a risk factor for that disorder, although the association may be weaker than that between the fixed marker and an incidence. Fixed markers are crucial to identification of “high risk” subpopulations, thus both to designing research studies and to targeting interventions, but since fixed markers cannot be changed, they are of little value to designing interventions to prevent onset of disorders.
The ideal, but admittedly unrealistic, approach to demonstrate risk factor status would be to sample the relevant population disorder free at a designated time zero (t = 0), to evaluate potential risk factors at that time, to follow each sampled individual over his/her subsequent lifetime, in order to evaluate and compare survival curves to the onset of the disorder [7] over that time period to see for whom and when onsets occur. With a binary risk factor (RF = 1 identifying “high risk” individuals and RF = 0 “low-risk”), one might see a situation such as that in Figure 1, which shows hypothetical survival curves to onset of a disorder for the “high risk” subpopulation (S1(t) in those with RF = 1) and the “low risk” subpopulation (S0(t) in those with RF = 0) for all values of t.
Then one could not only compare the overall survival curves, but could compare the incidence between time 0 and any fixed time T [1-S1(T) vs. 1-S0(T)]. In this hypothetical example, (1) the survival curves cross at t = T*, an unusual, although not an unknown, situation, and (2) onset is not inevitable for every individual, which is a quite common situation. In this example, 50% in one group have onset during their lifetimes compared to 30% in the other group, but the latter are likely to have their onset early if at all, which results in the crossing of survival curves, here at T* ≈ 18.
Here if incidences prior to T* are compared, RF would appropriately be described as a risk factor for the disorder; after T*, as a protective factor against the disorder. Exactly at T* the factor is unrelated to the incidence of the disorder. This illustrates the general principle that conclusions concerning the relationship of any risk factor for a disorder may depend strongly on the exact time span to which incidence refers.
Because it is more convenient to have short follow-up times, epidemiologists often assume that whatever the relative positions of the two survival curves near zero are the relative positions for all follow-up times. This is one of those extrapolations that often mislead both subsequent research efforts and clinical decision making. In general, the inferences from a cohort study apply to those in the disorder-free subpopulation represented in the sample at time 0, and followed for as long as the study chooses to follow the individuals, whether that be 1, 2, 5 or 10 years, but not necessarily to another population, and for any longer. Explicit presentation of the estimated survival curves up to the end of follow-up, as a general practice, would not only inform medical consumers as to exactly the findings of the study up to the duration of follow-up, but also remind them of the time limitations on inferences.
Standard survival methods [8] used to estimate survival curves deal well with “censored” data. Thus if the intended follow-up time is 10 years, but some individuals are lost to follow-up, still onset-free, after 1,2,3,… years, the survival curves can still be estimated up to 10 years using these methods, provided loss to follow-up is not associated with the same mechanism that produces onset. Thus, for example, in comparing those who did or did not use hormone replacement therapy (HRT) on time to onset of coronary disease [9], one cannot treat those with onset of cancer, cognitive disability, or the occurrence of a stroke as a “censored” data point, for HRT may be a risk factor for all these outcomes [10]. If such competing outcomes were mutually exclusive, observing the occurrence of any one before the others would indicate non-occurrence of the others at any later time, in which case that would not be a “censored” data point, but a signal of the longest possible survival time. However, such outcomes are not usually mutually exclusive. One may, for example, have both onset of coronary disease and subsequently onset of cognitive disability or vice versa. This is an unavoidable problem when risk factors are, as they often are, non-specific to one disorder, and epidemiologists seek to estimate the effect on each specific disorder separately.
Moreover, how the zero-point of time (t = 0) is defined makes a major difference in conclusions. For example, in the hypothetical situation in Figure 1, let us say that t = 0 refers to time of birth for the individuals in the population. If instead, individuals were sampled at the age of 10 or 20 or 50 (t = 0 in each case defined as the age at which individuals are sampled and their risk factors assessed), the populations sampled would differ, for those still disorder-free at age 10 are a non-random subset of those still disorder-free at age 20, etc. Moreover, any risk factor not a fixed marker might change within individuals from age 10 to 20 to 50, and the risk factor measured at different entry ages may have a different association with subsequent onset (age may moderate the effect of the risk factor on outcome). In general, the survival curves for those entering at different ages would be quite different from each other, and the conclusions might differ as well.
More problematic is the situation in which disorder-free individuals over a wide span of ages, say 20–80 years, are sampled, and t = 0 refers to the more or less arbitrary time each individual enters the research study. Time of entry to an observational research study (here the focus) has no clinical relevance to the individuals in the population to which inferences are to be drawn. In that case, the observed survival curve is a mixture of the survival curves of those who enter at each age disorder-free with the risk factors as they are at that age, the mixture determined by the age distribution in the samples. Different studies are unlikely to reproduce and confirm the same findings, particularly when both the factor (e.g., use of HRT) and the disorder itself (e.g., heart disease, diabetes, cancer, Alzheimer’s disease) are age-related.
Again, there is one rare situation that is an exception to these concerns: the constant hazards situation with fixed markers. If the risk factor is constant over the lifetime of the individual and the probability of survival for any time span is exactly the same regardless of when a individual enters a study (exponential survival curve, Poisson distribution of events), it doesn’t matter at what times the individuals are sampled, or whether they are sampled at the same time or what the distribution of entry times in the high- and low- risk subpopulations were. It doesn’t even matter how long individuals are followed or why they drop out. This is also the one and only case in which the incidence rate (“events per person-year”) estimates an interpretable population parameter, namely the reciprocal of the mean time to event, regardless of entry and exit times [11].
However, not only are many risk factors of interest not fixed (e.g., HRT use), there are few, if any, real onset distributions that follow a constant hazards model. Only an inevitable outcome (e.g., death) can possibly follow a constant hazards model. No age-related disorder (e.g., heart disease, cancer, Alzheimer’s disease or even death) can. Nevertheless epidemiological studies still occasionally use the incidence rate to compare the high- and low-risk subgroups [9,12] as well as other methods that depend on an unacknowledged assumption of constant hazards. Such studies have a high chance of drawing misleading conclusions [3].
In a prospective observational study, with a representative sample from the disorder-free population of interest, all entered at a relevant zero time (fixed short age span or a fixed life event) and followed for a reasonable fixed period of time, the association between a risk factor and onset is due either to a causal effect of the risk factor, and/or to any factor(s) associated with that risk factor. Attribution of causality to risk factors detected in observational studies must be tentative. In what follows, because the focus is on observational studies, no causal inferences will be drawn. Inferences apply only to the population sampled free of the disorder at a relevant zero time. For clarity of communication, all risk factors considered here are binary (RF = 1 vs. RF = 0), although the principles apply more generally. Finally, the outcome of interest is incidence between time zero at which time RF is measured, and a fixed time T for all individuals, the object to compare 1-S1(T) vs. 1-S0(T).

3. Statistical Significance is NOT the Goal: Public Health Significance Is

Over the last 20 years, considerable attention has been paid to the overuse, misuse, abuse of “statistical significance” [1321]. In the way statistical hypothesis testing is usually used, to say an association between a risk factor and subsequent onset of disorder is statistically significant is at best to say that the sample was large enough to detect a non-random effect. Such a non-random effect may be of trivial public health significance. What is usually reported is the “p-value”, a statistic computed with the study data. A level of significance is set ‘a priori’, α (typically 0.05). If p < α, then the result is said to be “statistically significant” at the α level of significance.
However, unless the null hypothesis is absolutely true, the expected value of the p-value approaches zero as the sample size increases, rapidly for a strong effect, slowly for an effect of trivial public health significance. Meehl and others [22,23] point out that the null hypothesis of randomness is never absolutely true. If so, no matter how trivial the non-random association between risk factor and outcome, it will be found “statistically significant” at whatever significance level is specified, provided only that the sample size is large enough. Consequently, the p-value serves more as an indicator of adequacy of the sample size to detect whatever effect size actually exists, than as an indicator of public health significance. A reviewer suggests, however, that there is a danger in ignoring the chance element in determining the p-value, that interpreting the p-value as equivalent to “adequate power” could lead to unacceptable publication bias. To show public health significance, an interpretable effect size is necessary, a population parameter that reflects the strength of association between the risk factor and the onset for public health purposes. There are a number of viable such effect sizes.
For example:
  • AUC [2429]is the probability that a low-risk individual will have better outcome than a high-risk one, where ties are broken with a toss of a fair coin: here AUC = .5(S0(T) − S1(T) + 1).
  • Success Rate Difference (SRD) [25,30] is the difference between the probability that a low-risk individual will have better outcome than a high-risk one and the probability that a high-risk individual will have a better outcome than a low-risk one: here SRD = S0(T) − S1(T) = 2AUC − 1 (in epidemiology, SRD is usually called the risk difference).
  • Number Needed to Take (NNT) [25,3138] equals 1/SRD or 1/(2AUC - 1). For a binary risk factor and incidence between 0 and T, NNT is the number of high risk individuals one would have to take at time 0 to find one more onset prior to T than if the same number of low risk individuals had been taken.
While these are three mathematically equivalent effect sizes, generally NNT is easier to interpret in terms of public health significance, and SRD and AUC are easier used in computations (e.g., for confidence intervals).
There are, of course, other viable effect sizes applicable in special circumstances. For example, Cohen’s d [39] is appropriate when the survival times in the high- and low- risk groups are normally distributed (an unusual situation): d = (μ1 − μ0)/σ, where μ1 and μ0 are the two group mean times of onset, and σ2 is the average of the two group variances. Then SRD = 2Φ(d/√2) − 1, where Φ() is the cumulative standard normal distribution. In the rare situation in which the constant hazards model holds in both groups, SRD = (μ1 − μ2)/(μ1 + μ2) = (RR − 1)/(RR + 1), where RR, a relative risk, is the ratio of the incidence rates in the two groups. There is limited applicability of such specialized effect sizes, but, when applicable, they are easily converted to SRD (NNT, AUC).
While the expected value of the p-value approaches zero as sample size increases, the sample estimate of an effect size, e.g., SRD, estimates the same population parameter regardless of the sample size. Instead, as sample size increases, the width of its confidence interval decreases to zero, i.e., large sample size improves the accuracy of effect size estimation, and does not change the effect size estimated.
A question that deserves careful future consideration is which values of NNT indicate public health significance, and which are trivial. For example, one would question any recommendation for costly and risky surgery on 500 patients to prevent one onset of coronary disease, i.e., to subject 499 patients unnecessarily to costly and risky surgery to prevent one onset. On the other hand, one might well be willing to recommend half a baby aspirin a day to 500 patients to prevent one such onset. In short, there is no universal answer to this question—the answer depends on how serious the disorder is, the consequences of untreated disorder, whether effective preventive intervention is available, how costly and risky such intervention might be, the vulnerability of the population and other such considerations. With the focus to date on “statistically significant” results, such questions have yet to be discussed seriously. Instead larger and larger sample sizes are used to detect ever smaller effect sizes, many of trivial public-health significance.
Epidemiologists often use the Odds Ratio (OR) as such an effect size, but OR is not viable in this role. Historically, OR was introduced as the likelihood-ratio test statistic to test the null hypothesis of randomness. OR remains useful as a detector of non-randomness, for example, in logistic regression analysis models: OR equal to 1 indicates random association; greater than 1, positive association, and less than 1, negative association. However, there is no magnitude of Odds Ratio unequal to 1 that unequivocally indicates public health significance.
Many arguments have been put forward in recent years to support that contentious point [4044], but let us here consider only one closely related to the consideration of time. In Figure 2 is shown the ROC (Receiver Operating Characteristic) plane in which the two survival curves shown in Figure 1 are compared. Here 1-S1(t) is graphed against 1-S0(t) for all values of t. These values connected with each other and with the two endpoints at (0,0) and (1,1) form the ROC curve comparing the RF = 1 and RF = 0 groups on time to onset (AUC is the area under this ROC curve). If there were only random association between the risk factor and onset, the ROC curve would coincide with the diagonal line from (0,0) to (1,1): the Random ROC. Here the ROC curve clearly indicates non-random association. The ROC curve crosses the Random ROC when t = T*, and as t increases, all points converge to the single point (0.5,0.3) determined by the proportion of the two groups who will eventually have this non-inevitable onset.
For any fixed follow-up time, T, there is one point on the ROC curve (1-S1(T), 1-S0(T)) that indicates the strength of association between the risk factor and that particular incidence. The SRD for such a binary outcome is proportional to the distance between that point and the Random ROC[42]. Since all ROC curves begin in the lower corner of the ROC plane, it is clear that the SRD measuring the association between the risk factor and any incidence depends strongly on T, and will always approach SRD = 0 as T approaches zero. However, Odds Ratio often tends to go in the opposite direction, becoming very large for very short follow-up periods, in some cases even approaching infinity as T approaches zero.
To demonstrate this and to understand why this is often so, it is necessary to put OR and SRD on comparable scales. In Figure 3 is shown the location of all pairs of probabilities (p1, p0) that would yield OR = 4 (OR = p1(1 − p0)/((1 − p1)p0)), and all pairs of probabilities that would yield SRD = 1/3 or NNT = 1/SRD = 3 (equipotency curves[43]). The same demonstration could be done for any fixed value of OR > 1, with SRD = (OR1/2 − 1)/(OR1/2 + 1) = Y (Yule’s Index).
What is in Figure 3 seen is generally true. SRD is constant (here equal 3) for all pairs of probabilities a fixed distance above the Random ROC. On the other hand, OR is constant (here equal 4) for all values on a curve symmetric around the line p1 = 1 − p0, and beginning and ending at the points of the Random ROC at (0,0) and (1,1). For fixed OR > 1, the maximal distance from the Random ROC coincides with the constant distance for SRD = (OR1/2 − 1)/(OR1/2 + 1). That distance then decreases to zero at both corners of the ROC plane. Thus it is always true that for positive association, SRD ≤ (OR1/2 − 1)/(OR1/2 + 1). In the special case when p1 = 1 − p0, the SRD = Y = (OR1/2 − 1)/(OR1/2 + 1). When p0 and p1 are both of moderate size (say between 0.25 and 0.75) then SRD is approximately equal to Y = (OR1/2 − 1)/(OR1/2 + 1).
In a ROC comparing survival curves as in Figure 3, to be informed, for example, that NNT = 3 (SRD = 1/3) is to be assured that the location of the point (1 − S1(T), 1 − S0(T)) is bounded away from random association. To be informed that OR = 4 allows the possibility of being as far away from random as is NNT = 3, but also allows the possibility of being arbitrarily close to random association, particularly when T is small.
In the comparison of any two survival curves, points corresponding to very short follow-up times are always near the lower left corner of the ROC plane, and very near the Random ROC. These points will have SRD = 1/NNT near zero indicating weak association, but since all the OR > 1 equipotency curves converge in that corner, these points often have very large OR. Simply stated, the problem is that the denominator of Odds Ratio approaches zero as T approaches zero, and division by zero tends both to “explode” the magnitude of any ratio and to make it very unstable. For this reason and all its many consequences, Odds Ratio should continue to be used as an indicator of non-randomness, to test null hypotheses of randomness, but not to be used as an effect size.
In Figure 4 are shown SRD = 1/NNT and Y = (OR1/2 − 1)/(OR1/2 + 1) for all follow-up times T, for the two survival curves in Figures 1,2. It can there be seen that for T near zero, OR is very large (Y = 0.4 corresponds to OR = 5.4), but SRD is near zero (NNT approaching infinity). For t = T*, as appropriate, both OR = 1 and SRD = 0 indicate random association. For t > T*, when both survival probabilities are in the middle range between 0.25 and 0.75, SRD is approximately equal to Y = (OR1/2 − 1)/(OR1/2 + 1).

4. How Risk Factors “Work Together”

Since the common use of Odds Ratio as if it were an effect size tends to exaggerate the association between risk factors and disorders, moving to use of SRD, NNT or AUC instead will only diminish the apparent clinical importance of many risk factors. This is unwelcome news, but probably reflective of the truth. A few special cases such as infectious diseases or single gene disorders aside, there are probably very few disorders for which a single risk factor can completely explain onset. It is likely that for complex disorders (heart disease, cancer, psychiatric disorders) multiple risk factors “work together” in parallel or in sequence to have influence the onset of a disorder. Thus examining how risk factors “work together” is crucial to prediction and prevention efforts.
To date, multiple possible risk factors are often simply included as independent variables in a linear model, completely ignoring (1) their timing relative to each other, (2) possible correlations between risk factors, (3) their possible interactive effects on incidence. Moreover the linearity assumptions and the link function selected (e.g., log-odds) often do not well fit the population.
The MacArthur Model is an alternative approach that takes these factors into consideration [4549]. The distribution of two binary risk factors RF1 and RF2 in the population of interest is shown in Table 1. In that population the probabilities that RF1 = 1 is Q, and that RF2 = 1 is P. The parameter ρ (the product moment correlation or phi coefficient between the risk factors) is an indicator of non-random association between the two risk factors in that population, with ρ = 0 indicating stochastic independence between them.
In Table 2 are shown the incidences between time 0 and fixed T, in each of the four subgroups defined by the two risk factors. When using SRDs as the effect size, the conditional SRDs for RF1 for the two values of RF2, and the conditional SRDs for RF2 for the two values of RF1 are also shown. The “main effect of RF1” (ME1), and the “main effect of RF2” (ME2) are respectively the averages of the corresponding conditional SRDs, and the “interaction effect of RF1 and RF2” (INT) is the difference between those conditional SRDs (the same for both sets of conditional SRDs).
If RF2 were ignored, the SRD relating RF1 to outcome (for each possible value of T) is equal to:
SRD 1 = ME 1 + ρ ( PP ' / QQ ' ) 1 / 2 ME 2 + .5 * INT [ ( 2 P 1 ) ρ ( PP ' / QQ ' ) 1 / 2 ( 2 Q 1 ) ] .
If RF1 were ignored, the SRD relating RF2 to outcome is equal to:
SRD 2 = ME 2 + ρ ( QQ ' / PP ' ) 1 / 2 ME 1 + .5 * INT [ ( 2 Q 1 ) ρ ( QQ ' / PP ' ) 1 / 2 ( 2 P 1 ) ] .
SRD1 and SRD2 are called the “raw”, “overall”, “marginal”, or “univariate” effects of RF1 and RF2 on the outcome, indicating the association of that risk factor on outcome in the population sampled when all other risk factors are ignored. While the formulas above are exact only for two binary risk factors predicting incidence between 0 and fixed T, the principle is true in general. The “raw” effect of a risk factor (SRD1 or SRD2) comprises three sources: the unique effect of the risk factor itself, the main effects of other risk factors correlated with the risk factor of interest, and the interactions of the risk factor of interest with other risk factors. How much of each source is represented depends on the joint distributions of the risk factors in the population (here P and Q and their correlation as indicated in Table 1).
The main effect of a risk factor of interest (ME1 or ME2) does not convey the effect size in the total population unless the other risk factor(s) are neither correlated (ρ = 0) nor interactive (INT = 0) for the risk factor of interest. Nor does the main effect of a risk factor convey the strength of association in each subpopulation “matched” on other risk factor, unless there is no interactive effect. In short, the research questions addressed by the raw effect size SRD1, the conditional SRDs for risk factor 1 in the subpopulations with RF2 = 1 and 0, and the main effect of RF1, are all usually different, not because one is “right” and the others “wrong”, but because they address the association of RF1 with outcome in different populations. Thus “adjusting” for RF2 in a linear model does not usually “remove the effect of RF2”. It changes the research question from that of focusing on the effect size of RF1 in the total population sampled, to that of the common effect size of RF1 in the subpopulations defined by RF2 in absence of an interaction, or to some weighting of those effect sizes in the presence of an interaction, the weighting determined by whether or not the interaction was included in the linear model and the joint distribution of the risk factors.
When risk factors are coded +/−1/2 in a linear model (chosen because SRD is here the effect size):
Main   Effect   of   RF 1 : [ S 10 ( T ) S 11 ( T ) + S 00 ( T ) S 01 ( T ) ] / 2 Main   Effect   of   RF 2 : [ S 01 ( T ) S 11 ( T ) + S 00 ( T ) S 10 ( T ) ] / 2 Interactive   Effect : [ S 10 ( T ) S 11 ( T ) S 00 ( T ) + S 01 ( T ) ]
If there is no temporal precedence for RF1 and RF2, then there are three possible roles for RF1 and RF2 for the incidence in question:
RF1 and RF2 are “independent risk factors” if ρ = 0, and if both matter, i.e., INT is non-zero or both ME1 and ME2 are non-zero. In such cases, the two risk factors play a joint role in determining the outcome and would continue to be of parallel interest. For many disorders, gender and ethnicity are independent risk factors.
One of RF1 or RF2 “is proxy to” the other if ρ is unequal to zero (RF1 and RF2 are correlated risk factors), and only one matters. Thus if ME1 = INT = 0, then RF1 is proxy to RF2. If ME2 = INT = 0, then RF2 is proxy to RF1. In such cases, the proxy variable should be set aside from further consideration. For example, a measure of family income is often proxy to a well-measured socio-economic index for the family, because family income is usually one component of the index, but less reliably measured than the index as a whole.
RF1 and RF2 are overlapping risk factors if ρ is unequal to zero, and both matter. Thus if any two of ME1, ME2 and INT are not equal to zero, RF1 and RF2 are overlapping. This situation often arises when two risk factors tap the same underlying construct with about equal, but less than perfect, reliability/validity. Then it is preferable to combine the two risk factors to generate a more reliable/valid measure of whatever their common construct. Combining two somewhat unreliable measures of the same construct would disattenuate reliability and thus increase the effect size. Moreover, to do so might focus attention more precisely on the appropriate underlying causal factor. For example, infant birth-weight and gestational age tend to be highly correlated and risk factors for many of the same subsequent outcomes. Some measure of birth maturity that included consideration of both, perhaps even including other indicators of physiological and neurological maturity at birth, might serve prediction and prevention purposes better than either separately.
On the other hand, when there is temporal precedence, with RF1 preceding RF2 in time, there are four possibilities:
RF1 moderates RF2 if ρ = 0 and INT is non-zero. Then the conditional effect sizes for RF2 differ depending on whether RF1 = 1 or = 0. Since a later risk factor, RF2, “works differently” depending on what earlier RF1 is, this suggests that the population should be stratified on RF1 for further studies. For example, a genotype may be a susceptibility factor for a later environmental risk factor. For those with one genotype, RF2 may be a strong risk factor for outcome; otherwise, RF2 may be a much weaker risk factor, may not be a risk factor at all, or may even be a protective factor. Indeed, seeking genetic moderators of drug on therapeutic response is the basis for current interest in pharmacogenetics. Moderators are also the basis of personalized medicine[5053]
RF2 mediates RF1 if ρ is non-zero, and either INT or ME2 is non-zero. In this case, RF2 explains part of the effect of earlier RF1 on the outcome. When a mediator is identified, this suggests the possibility of a chain leading from RF1 through RF2 to the outcome. For example, unsafe sex practices lead to HIV infection that leads to onset of AIDS: HIV infection mediates the effect of unsafe sex practices on AIDS. Mediator relationships are important in that the chain provides multiple opportunities for preventive intervention: one might break the chain by breaking any of the links in the chain.
RF2 is proxy to RF1 if ρ is non-zero, and if MS2 and INT are both zero. As is the case for proxy risk factors in absence of temporal precedence, the proxy factor should be set aside. Gender, for example is a risk factor for teen onset of depression. There are many correlates of gender during the pre-teen years, e.g., ball-throwing ability at age 10, that might be found to be risk factors for teen-onset depression when considered individually, but would be found proxy to gender when both were considered. It is probably not worthwhile to teach young girls to throw a ball better to prevent teen depression.
RF1 and RF2 are independent risk factors if ρ = 0, but INT = 0, i.e., RF1 does not moderate RF2. As is the case for independent risk factors in absence of temporal precedence, both factors would continue to be of parallel interest.
It should be noted that these definitions are more precise than certain current usages in epidemiology. For example, “independent risk factors” in the MacArthur model are required to be stochastically independent of each other. Usual usage of the term does not require such independence. In the way the term is often used, only proxy risk factors would not be labeled independent risk factors, given large enough sample sizes.
The Last[1] definitions of “mediator” (or “mediating variable” or “intermediate variable”) and of “moderator variable” (or “qualifier variable”) require that the causal pathway from the independent (risk factor of interest) to the dependent variable (onset) be known. Such information is generally not available to a cohort study. Here the issues are, first, temporal precedence and correlation and then the joint association of two risk factors with the outcome. Causality is neither assumed nor inferred.
Moreover the term “confounder” is avoided. Last defines the term as “A variable that can cause or prevent the outcome of interest, is not an intermediate variable (mediator), and is associated with the factor under investigation.” (Page 35). That would preclude mediators explicitly, and preclude moderators and independent risk factors because they are not associated with the factor under investigation, leaving proxies or overlapping factors. However, in practice, the term “confounder” is often used more loosely to refer to risk factors in which the researcher is not specifically interested. Thus in a study examining the relationship of diet and exercise to onset of obesity, a dietician might designate exercise as a “confounder”, while an exercise physiologist might designate diet as a “confounder”.
There is as yet little history of seeking moderating/mediating relationships between risk factors for specific outcomes. Consequently, how to conduct such a search to general moderator/mediator hypothesis, and how to conduct studies to formally test such hypotheses, is still work in progress. However, there are a few examples in the literature that might suggest possible options [5460].

5. Discussion

Many of the problems here discussed are long and well-known, and yet continue to occur. For example, Caspi and colleagues [57] reported in 2003 on the moderating effect of a gene (5-HTT) on an environmental risk factor (stress) on major depression (DSM-IV definition) in a prospective community cohort followed from birth to 26 years. Their finding was particularly interesting since it suggests a genetic moderator of an environmental effect on a disorder. Risch et al. in 2009 [60] evaluated whether that finding had since been confirmed or refuted. However, of the 13 studies they reported as attempts to replicate, 8 were cross-sectional studies, that could legitimately evaluate neither stress as a risk factor for major depression, nor any moderator hypothesis. Of the five remaining prospective cohort studies, none covered the same age range as did the Caspi et al. study, with two sampling those 65 years of age or older. No effect sizes other than Odds Ratios were reported for any of the studies. Thus, none of the reported studies actually evaluated the same effect as did the Caspi et al. study, largely because issues related to time were ignored.
Science progresses by identifying its weaknesses and repairing them. However, it is always very hard to give up methods long used in previous studies and thus very familiar, to be replaced with new and unfamiliar methods. “That’s not the way it is always done, or the way everyone does it!” is a common rejoinder to suggestions for alternative approaches to deal with the problems here discussed.
Such resistance is particularly and predictably strong when the alternative approaches are more difficult and costly to implement, which is here true. To be asked to stratify a mixed age sample, say 20–80 years of age, into relatively short age strata (say entry at 20–24, 25–29, etc.), and to do analysis separately within each age stratum, results in a reduction of sample size and thus power to detect effects, inevitably an unwelcome suggestion. However, the alternative is the risk of detection of false positive results.
Resistance is particularly, and again, understandably, strong when alternative methods are less likely to produce “significant” results, and more likely to indicate that the effect of any single risk factor is quite weak, because such alternative methods counter the tendency of incorrect methods to produce false positive results. This may be one of the major reasons why Odds Ratio has been so hard to displace as an effect size: it tends to exaggerate what are often trivial associations and results in publications of results that only later are shown to be exaggerations at best, and at worst, false positive results.
The rejoinder hardest to refute is that the effect of time can be dealt with simply by “adjusting for time” in a linear model. In some cases this may indeed be possible. However, the assumptions and fit of such models should be carefully checked, for if the assumptions that underlie valid results from application of such models do not hold in the population sampled, the results of such adjustments may be more biased than the results in absence of adjustment.
In summary, the issue of time should be central to all thinking in epidemiology research, which would necessitate careful thinking about sampling, measurement, design, analysis, and, perhaps most important, about the interpretation of the results from such studies that might influence clinical decision-making and subsequent clinical research.

References

  1. Last, JM. A Dictionary of Epidemiology; Oxford University Press: New York, NY, USA, 1995; p. 35. [Google Scholar]
  2. Ioannidis, JPA. Why most published research findings are false. PLoS Med 2005, 2, 696–791. [Google Scholar]
  3. Ioannidis, JPA. Contradicted and initially stronger effects in highly cited clinical research. J. Am. Med. Assn 2005, 294, 218–228. [Google Scholar]
  4. von Elm, R; Egger, M. The scandal of poor epidemiological research. Brit. Med. J 2004, 329, 868–869. [Google Scholar]
  5. Martyn, C. What is the point of epidemiology? Pract. Neurol 2004, 4, 192–193. [Google Scholar]
  6. Kraemer, HC; Kazdin, AE; Offord, DR; Kessler, RC; Jensen, PS; Kupfer, DJ. Coming to terms with the terms of risk. Arch. Gen. Psychiatr 1997, 54, 337–343. [Google Scholar]
  7. Singer, JD; Willett, JB. It’s about time: using discrete-time survival analysis to study duration and the timing of events. J. Educ. Stat 1993, 18, 155–195. [Google Scholar]
  8. Kaplan, EL; Meier, P. Nonparametric estimation from incomplete observations. J. Am. Stat. Assn 1958, 53. [Google Scholar]
  9. Stampfer, MJ; Colditz, GA; Willett, WC; Manson, JE; Rosner, B; Speizer, FE; Hennekens, CH. Postmenopausal estrogen therapy and cardiovascular disease. Ten-year follow-up from the nurses’ health study. N. Engl. J. Med 1991, 325, 756–762. [Google Scholar]
  10. Writing Group for the Women’s Health Initiative Investigators. Principal results from the Women’s Health Initiative randomized controlled trial. J. Am. Med. Assn 2002, 288, 321–333. [Google Scholar]
  11. Kraemer, HC. Events per person-time (incidence rate): a misleading statistic? Stat.Med 2009, 29, 1028–1039. [Google Scholar]
  12. Stampfer, MJ; Willett, WC; Colditz, GA; Rosner, B; Speizer, FE; Hennekens, CH. A prospective study of postmenopausal estrogen therapy and coronary heart disease. N. Engl. J. Med 1985, 313, 1044–1049. [Google Scholar]
  13. Kline, RB. Beyond Significance Testing: Reforming Data Analysis Methods in Behavioral Research; American Psychological Association: Washington, DC, WA, USA, 2005. [Google Scholar]
  14. Nickerson, RS. Null hypothesis significance testing: a review of an old and continuing controversy. Psychol. Methods 2000, 5, 241–301. [Google Scholar]
  15. Wilkinson, L. The task force on statistical inference statistical methods in psychology journals: guidelines and explanations. Am. Psychol 1999, 54, 594–604. [Google Scholar]
  16. Thompson, B. Journal editorial policies regarding statistical significance tests: heat is to fire as p is to importance. Educ. Psychol. Rev 1999, 11, 157–169. [Google Scholar]
  17. Krantz, DH. The null hypothesis testing controversy in psychology. J. Am. Stat. Assn 1999, 44, 1372–1381. [Google Scholar]
  18. Shrout, PE. Should significance tests be banned? Introduction to a special section exploring the pros and cons. Psychol. Sci 1997, 8, 1–2. [Google Scholar]
  19. Hunter, JE. Needed: A ban on the significance test. Psychol. Sci 1997, 8, 3–7. [Google Scholar]
  20. Cohen, J. The Earth is Round (p < 0.05). Am. Psychol 1995, 49, 997–1003. [Google Scholar]
  21. Dar, R; Serlin, RC; Omer, H. Misuse of statistical tests in three decades of psychotherapy research. J. Consult. Clin. Res 1994, 62, 75–82. [Google Scholar]
  22. Meehl, PE. Theoretical risks and tabular asterisks: Sir Karl, Sir Ronald, and the slow progress of soft psychology. J. Consult. Clin. Psychol 1978, 46, 806–834. [Google Scholar]
  23. Jones, LV; Tukey, JW. A sensible formulation of the significance test. Psychol. Methods 2000, 5, 411–414. [Google Scholar]
  24. Newcombe, RG. Confidence intervals for an effect size measure based on the Mann-Whitney statistic. Part 1: General issues and tail area based methods. Stat. Med 2006, 25, 543–557. [Google Scholar]
  25. Kraemer, HC; Kupfer, DJ. Size of treatment effects and their importance to clinical research and practice. Biol. Psychiatr 2006, 59, 990–996. [Google Scholar]
  26. Acion, L; Peterson, JJ; Temple, S; Arndt, S. Probabilistic index: an intuitive non-parametric approach to measuring the size of treatment effects. Stat. Med 2006, 25, 591–602. [Google Scholar]
  27. Grissom, RJ; Kim, JJ. Review of assumptions and problems in the appropriate conceptualization of effect size. Psychol. Methods 2001, 6, 135–146. [Google Scholar]
  28. Grissom, RJ. Probability of the superior outcome of one treatment over another. J. App. Psychol 1994, 79, 314–316. [Google Scholar]
  29. McGraw, KO; Wong, SP. A common language effect size statistic. Psychol. Bull 1992, 111, 361–365. [Google Scholar]
  30. Hsu, LM. Biases of success rate differences shown in binomial effect size displays. Psychol. Bull 2004, 9, 183–197. [Google Scholar]
  31. Wen, L; Badgett, R; Cornell, J. Number needed to treat: a descriptor for weighing therapeutic options. Am. J. Health-Syst. Pharm 2005, 62, 2031–2036. [Google Scholar]
  32. Bogarty, P; Brophy, J. Numbers needed to treat (needlessly?). The Lancet 2005, 365, 1307–1308. [Google Scholar]
  33. Altman, DG; Schulz, KF; Hoher, D; Egger, M; Davidoff, F; Elbourne, D; Gøtzsche, PC; Lang, T; Consort_Group. The revised CONSORT statement for reporting randomized trials: explanation and elaboration. Ann. Intern. Med 2001, 134, 663–694. [Google Scholar]
  34. Newcombe, RG. Confidence intervals for the number needed to treat-absolute risk reduction is less likely to be misunderstood. Bri. Med. J 1999, 318, 1765. [Google Scholar]
  35. Lesaffre, E; Pledger, G. A note on the number needed to treat. Contr. Clin. Trial 1999, 20, 439–447. [Google Scholar]
  36. Cook, RJ; Sackett, DL. The number needed to treat: a clinically useful measure of treatment effect. Brit. Med. J 1995, 310, 452–454. [Google Scholar]
  37. Altman, DG; Andersen, K. Calculating the number needed to treat for trials where the outcome is time to an event. Brit. Med. J 1999, 319, 1492–1495. [Google Scholar]
  38. Altman, DG. Confidence intervals for the number needed to treat. Brit. Med. J 1998, 317, 1309–1312. [Google Scholar]
  39. Cohen, J. Statistical Power Analysis for the Behavioral Sciences; Lawrence Erlbaum Associates: Hillsdale, NJ, USA, 1977. [Google Scholar]
  40. Kraemer, HC. Correlation coefficients in medical research: from product moment correlation to the odds ratio. Stat. Methods Med. Res 2007, 15, 525–545. [Google Scholar]
  41. Newcombe, RG. A deficiency of the odds ratio as a measure of effect size. Stat.Med 2006, 25, 4235–4240. [Google Scholar]
  42. Kraemer, HC. Reconsidering the odds ratio as a measure of 2X2 Association in a population. Stat. Med 2004, 23, 257–270. [Google Scholar]
  43. Kraemer, HC; Kazdin, AE; Offord, DR; Kessler, RC; Jensen, PS; Kupfer, DJ. Measuring the potency of a risk factor for clinical or policy significance. Psychol. Methods 1999, 4, 257–271. [Google Scholar]
  44. Sackett, DL. Down with odds ratios! Evid. Based Med 1996, 1, 164–166. [Google Scholar]
  45. Kraemer, HC; Stice, E; Kazdin, A; Kupfer, D. How do risk factors work together to produce an outcome? Mediators, moderators, independent, overlapping and proxy risk factors. Am. J. Psychiatr 2001, 158, 848–856. [Google Scholar]
  46. Kraemer, HC; Wilson, GT; Fairburn, CG; Agras, WS. Mediators and moderators of treatment effects in randomized clinical trials. Arch. Gen. Psychiatr 2002, 59, 877–883. [Google Scholar]
  47. Kraemer, HC; Frank, E; Kupfer, DJ. Moderators of treatment outcomes: clinical, research, and policy importance. J.Am. Med. Assn 2006, 296, 1–4. [Google Scholar]
  48. Kraemer, HC; Kiernan, M; Essex, MJ; Kupfer, DJ. How and why criteria defining moderators and mediators differ between the Baron & Kenny and MacArthur approaches. Health Psychol 2008, 27, S101–S108. [Google Scholar]
  49. Kraemer, HC. Toward non-parametric and clinically meaningful moderators and mediators. Stat. Med 2008, 27, 1679–1692. [Google Scholar]
  50. Jain, KK. Personalized medicine. Curr. Opin. Mol. Ther 2002, 4, 548–558. [Google Scholar]
  51. Abrahams, E; Ginsburg, GS; Silver, M. The personalized medicine coalition: goal and strategies. A. J. Pharmacogenomics 2005, 5, 345–355. [Google Scholar]
  52. Lesko, LJ. Personalized medicine: elusive dream or imminent reality? Cl. Pharmacol. Ther 2007, 81, 807–815. [Google Scholar]
  53. Richmond, TD. The current status and future potential of personalized diagnostics: streamlining a customized process. Biotechnol. Ann. Rev 2008, 14, 411–422. [Google Scholar]
  54. The MTA Cooperative Group. Moderators and mediators of treatment response for children with attention-deficit/hyperactivity disorder. J. Consult. Clin. Psychol 1999, 56, 1088–1096. [Google Scholar]
  55. Kraemer, HC; Wilson, GT; Fairburn, CG; Agras, WS. Mediators and moderators of treatment effects in randomized clinical trials. Arch. Gen. Psychiatr 2002, 59, 877–883. [Google Scholar]
  56. Owens, EG; Hinshaw, SP; Kraemer, HC; Arnold, LE; Abikoff, HB; Cantwell, DP; Conners, CK; Elliot, G; Greenhill, LL; Hechtman, L; Hoza, B; Jensen, PS; March, JS; Newcorn, JH; Pelham, WE; Richters, JE; Schiller, EP; Severe, JG; Swanson, JM; Vereen, D; Vitiello, B; Wells, KC; Wigal, T. Moderators of treatment response in the MTA. J. Consult. Clin. Psychol 2003, 71, 540–552. [Google Scholar]
  57. Caspi, A; Sugden, K; Moffitt, TE; Taylor, A; Craig, IW; Harrington, H; McClay, J; Mill, J; Martin, J; Braithwaite, A; Poulton, R. Influence of life stress on depression: moderation by a polymorphism in the 5-HTT gene. Science 2003, 301, 386–389. [Google Scholar]
  58. Boyce, WT; Essex, MJ; Alkon, A; Goldsmith, HH; Kraemer, HC; Kupfer, DJ. Early father involvement moderates biobehavioral susceptibility to mental health problems in middle childhood. J. Am. Acad. Child Adol. Psychiatr 2006, 45, 1510–1520. [Google Scholar]
  59. Essex, MJ; Kraemer, HC; Armstrong, JM; Boyce, WT; Goldsmith, HH; Klein, MH; Woodward, H; Kupfer, DJ. Exploring risk factors for the emergence of children’s mental health problems. Arch. Gen. Psychiatr 2006, 63, 1246–1256. [Google Scholar]
  60. Risch, N; Herrell, R; Lehner, T; Lian, K-Y; Eaves, L; Hoh, J; Griem, A; Kovacs, M; Ott, J; Merikangas, KR. Interactions between the serotonin transporter gene (5-HTTLPR), stressful life events, and risk of depression: a Meta-analysis. J. Am. Med. Assn 2009, 301, 2462–2471. [Google Scholar]
Figure 1. Comparison of two hypothetical survival curves for the subpopulations with RF = 1 and RF = 0 (non-proportional hazards).
Figure 1. Comparison of two hypothetical survival curves for the subpopulations with RF = 1 and RF = 0 (non-proportional hazards).
Ijerph 07 00029f1
Figure 2. ROC curve comparing survival in the “high” and “low” risk groups.
Figure 2. ROC curve comparing survival in the “high” and “low” risk groups.
Ijerph 07 00029f2
Figure 3. Equipotency curves comparing the locus of all points (p0, p1) with p1 > p0, the ROC plane with Odds Ratio = p1(1 − p0)/[(1 − p1)p0] and NNT = 1/(p1 − p0) = 3. (The Figure 3 here selected to be equal to (OR1/2 + 1)/(OR1/2 − 1)).
Figure 3. Equipotency curves comparing the locus of all points (p0, p1) with p1 > p0, the ROC plane with Odds Ratio = p1(1 − p0)/[(1 − p1)p0] and NNT = 1/(p1 − p0) = 3. (The Figure 3 here selected to be equal to (OR1/2 + 1)/(OR1/2 − 1)).
Ijerph 07 00029f3
Figure 4. Comparison of Y = (OR1/2 − 1)/(OR1/2 + 1) with 1/NNT for various follow-up times for the survival curves shown in Figure 1.
Figure 4. Comparison of Y = (OR1/2 − 1)/(OR1/2 + 1) with 1/NNT for various follow-up times for the survival curves shown in Figure 1.
Ijerph 07 00029f4
Table 1. The joint distribution of two binary risk factors (RF1 and RF2). with marginal probabilities P = Prob(RF2 = 1) and Q = Prob(RF1 = 1), and the product moment correlation coefficient (ρ) between them.
Table 1. The joint distribution of two binary risk factors (RF1 and RF2). with marginal probabilities P = Prob(RF2 = 1) and Q = Prob(RF1 = 1), and the product moment correlation coefficient (ρ) between them.
RF1 = 1RF1 = 0
RF2 = 1PQ+ρ(PP’QQ’)1/2PQ’−ρ(PP’QQ’)1/2P
RF2 = 0P’Q−ρ(PP’QQ’)1/2P’Q’+ ρ(PP’QQ’)1/2P’ = 1 − P
QQ’ = 1 − Q
Table 2. The incidence of disorder by time T for each combination of RF1 and RF2, and the marginal SRDs.
Table 2. The incidence of disorder by time T for each combination of RF1 and RF2, and the marginal SRDs.
RF1 = 1RF1 = 0Marginal SRD for RF1
RF2 = 11 − S11(T)1 − S10(T)S10(T)−S11(T)
RF2 = 01 − S01(T)1 − S00(T)S00(T)−S01(T)
Marginal SRD’s for RF2S01(T)−S11(T)S00(T)−S10(T)

Share and Cite

MDPI and ACS Style

Kraemer, H.C. Epidemiological Methods: About Time. Int. J. Environ. Res. Public Health 2010, 7, 29-45. https://doi.org/10.3390/ijerph7010029

AMA Style

Kraemer HC. Epidemiological Methods: About Time. International Journal of Environmental Research and Public Health. 2010; 7(1):29-45. https://doi.org/10.3390/ijerph7010029

Chicago/Turabian Style

Kraemer, Helena Chmura. 2010. "Epidemiological Methods: About Time" International Journal of Environmental Research and Public Health 7, no. 1: 29-45. https://doi.org/10.3390/ijerph7010029

Article Metrics

Back to TopTop