-
PDF
- Split View
-
Views
-
Cite
Cite
Ted Westling, Peter Gilbert, Marco Carone, Causal Isotonic Regression, Journal of the Royal Statistical Society Series B: Statistical Methodology, Volume 82, Issue 3, July 2020, Pages 719–747, https://doi.org/10.1111/rssb.12372
- Share Icon Share
Summary
In observational studies, potential confounders may distort the causal relationship between an exposure and an outcome. However, under some conditions, a causal dose–response curve can be recovered by using the G-computation formula. Most classical methods for estimating such curves when the exposure is continuous rely on restrictive parametric assumptions, which carry significant risk of model misspecification. Non-parametric estimation in this context is challenging because in a non-parametric model these curves cannot be estimated at regular rates. Many available non-parametric estimators are sensitive to the selection of certain tuning parameters, and performing valid inference with such estimators can be difficult. We propose a non-parametric estimator of a causal dose–response curve known to be monotone. We show that our proposed estimation procedure generalizes the classical least squares isotonic regression estimator of a monotone regression function. Specifically, it does not involve tuning parameters and is invariant to strictly monotone transformations of the exposure variable. We describe theoretical properties of our proposed estimator, including its irregular limit distribution and the potential for doubly robust inference. Furthermore, we illustrate its performance via numerical studies and use it to assess the relationship between body mass index and immune response in human immunodeficiency virus vaccine trials.
1 Introduction
1.1 Motivation and literature review
Questions regarding the causal effect of an exposure on an outcome are ubiquitous in science. If investigators can carry out an experimental study in which they randomly assign a level of exposure to each participant and then measure the outcome of interest, estimating a causal effect is generally straightforward. However, such studies are often not feasible, and data from observational studies must be relied on instead. Assessing causality is then more difficult, in large part because of potential confounding of the relationship between exposure and outcome. Many non-parametric methods have been proposed for drawing inference about a causal effect by using observational data when the exposure of interest is either binary or categorical—these include, among others, inverse-probability-weighted estimators (Horvitz and Thompson, 1952), augmented inverse-probability-weighted (AIPW) estimators (Scharfstein et al., 1999; Bang and Robins, 2005) and targeted minimum-loss-based estimators (van der Laan and Rose, 2011).
In practice, many exposures are continuous, in the sense that they may take any value in an interval. A common approach to dealing with such exposures is simply to discretize the interval into two or more regions, thus returning to the categorical exposure setting. However, it is frequently of scientific interest to learn the causal dose–response curve, which describes the causal relationship between the exposure and outcome across a continuum of the exposure. Much less attention has been paid to continuous exposures. Robins (2000) and Zhang et al. (2016) studied this problem by using parametric models, and Neugebauer and van der Laan (2007) considered inference on parameters obtained by projecting a causal dose–response curve onto a parametric working model. Other researchers have taken a non-parametric approach instead. Rubin and van der Laan (2006) and Díaz and van der Laan (2011) discussed non-parametric estimation using flexible data-adaptive algorithms. Kennedy et al. (2017) proposed an estimator based on local linear smoothing. Finally, van der Laan et al. (2018) have recently presented a general framework for inference on parameters that fail to be sufficiently smooth as a function of the data-generating distribution and for which regular root-n-estimation theory is therefore not available. This is indeed the case for the causal dose–response curve, and van der Laan et al. (2018) discussed inference on such a parameter as a particular example.
Despite a growing body of literature on non-parametric estimation of causal dose–response curves, to the best of our knowledge, existing methods do not permit valid large sample inference and may be sensitive to the selection of certain tuning parameters. For instance, smoothing-based methods are often sensitive to the choice of a kernel function and bandwidth, and these estimators typically have non-negligible asymptotic bias, which complicates the task of performing valid inference.
In many settings, it may be known that the causal dose–response curve is monotone in the exposure. For instance, exposures such as daily exercise performed, cigarettes smoked per week and air pollutant levels are all known to have monotone relationships with various health outcomes. In such cases, an extensive literature suggests that monotonicity may be leveraged to derive estimators with desirable properties—Groeneboom and Jongbloed (2014) provides a comprehensive overview. For example, in the absence of confounding, isotonic regression may be employed to estimate the causal dose–response curve (Barlow et al., 1972). The isotonic regression estimator does not require selection of a kernel function or bandwidth, is invariant to strictly increasing transformations of the exposure and on centring and scaling by n−1/3, converges in law pointwise to a symmetric limit distribution with mean 0 (Brunk, 1970). The convergence property is useful since it facilitates asymptotically valid pointwise inference.
Non-parametric inference on a monotone dose–response curve when the exposure–outcome relationship is confounded is more difficult to tackle and is the focus of this paper. To the best of our knowledge, this problem has not been comprehensively studied before.
1.2 Parameter of interest and its causal interpretation
The prototypical data unit that we consider is O = (Y, A, W), where Y is a response, A a continuous exposure and W a vector of covariates. The support of the true data-generating distribution is denoted by , where , is an interval and . Throughout, the use of subscript 0 refers to evaluation at or under . For example, we write and F0 to denote and respectively, and E0 to denote expectation under .
Suppose that
- (a)
each unit’s potential outcomes are independent of all other units’ exposures, and
- (b)
the observed outcome Y equals the potential outcome Y(A) corresponding to the exposure level A that is actually received.
Identification of m0(a) further depends on the relationship between A and Y(a). If assumptions (a) and (b) hold and, in addition,
- (a)
A and Y(a) are independent, and
- (b)
the marginal density of A is positive at a,
then m0(a) = r0(a). Condition (c) typically holds in experimental studies only (e.g. randomized trials). In observational studies there are often common causes of A and Y(a)—so-called confounders of the exposure–outcome relationship—that induce dependence. In such cases, m0(a) and r0(a) do not generally coincide. However, if W contains a sufficiently rich collection of confounders, it may still be possible to identify m0(a) from the observed data. If assumptions (a) and (b) hold and, in addition,
- (a)
A and Y(a) are conditionally independent given W, and
- (b)
the conditional density of A given W is almost surely positive at A = a,
then . This is a fundamental result in causal inference (Robins, 1986; Gill and Robins, 2001). Whenever , our methods can be interpreted as drawing inference on the causal dose–response parameter m0(a).
We note that the definition of the counterfactual outcome Y(a) presupposes that the intervention setting A = a is uniquely defined. In many situations, this stipulation requires careful thought. For example, in Section 6 we consider an application in which body mass index BMI is the exposure of interest. There is an on-going scientific debate about whether such an exposure leads to a meaningful causal interpretation, since it is not clear what it means to intervene on BMI.
Even if the identifiability conditions that were stipulated above do not strictly hold or the scientific question is not causal in nature, when W is associated with both A and Y, often has a more appealing interpretation than the unadjusted regression function r0(a). Specifically, may be interpreted as the average value of Y in a population with exposure fixed at A = a but otherwise characteristic of the study population with respect to W. Because involves both adjustment for W and marginalization with respect to a single reference population that does not depend on the value a, the comparison of over different values of a is generally more meaningful than for r0(a).
When , the parameter P↦θP(a) is not pathwise differentiable at with respect to the non-parametric model (Díaz and van der Laan, 2011). Heuristically, because of the continuous nature of A, θP(a) corresponds to a local feature of P. As a result, regular root-n-rate estimators cannot be expected, and standard methods for constructing efficient estimators of pathwise differentiable parameters in non-parametric and semiparametric models (e.g. estimating equations, one-step estimation or targeted minimum loss-based estimation) cannot be used directly to target and obtain inference on .
1.3 Contribution and organization of the paper
We denote by the distribution function of A under P, by the class of non-decreasing real-valued functions on and by the class of strictly increasing and continuous distribution functions supported on . The statistical model that we shall work in is , which consists of the collection of distributions for which θP is non-decreasing over and the marginal distribution of A is continuous with positive Lebesgue density over .
In this paper, we study non-parametric estimation and inference on the G-computed regression function for use when A is a continuous exposure and is known to be monotone. Specifically, our goal is to make inference about for by using independent observations O1, O2, …, On drawn from . This problem is an extension of classical isotonic regression to the setting in which the exposure–outcome relationship is confounded by recorded covariates—this is why we refer to the method proposed as causal isotonic regression. As mentioned above, to the best of our knowledge, non-parametric estimation and inference on a monotone G-computed regression function have not been comprehensively studied before. In what follows, we
- (a)
show that our proposed estimator generalizes the unadjusted isotonic regression estimator to the more realistic scenario in which there is confounding by recorded covariates,
- (b)
investigate finite sample and asymptotic properties of the estimator proposed, including invariance to strictly increasing transformations of the exposure, doubly robust consistency and doubly robust convergence in distribution to a non-degenerate limit,
- (c)
derive practical methods for constructing pointwise confidence intervals, including intervals that have valid doubly robust calibration and
- (d)
illustrate numerically the practical performance of the estimator.
We note that, in Westling and Carone (2019), we studied estimation of as one of several examples of a general approach to monotonicity-constrained inference. Here, we provide a comprehensive examination of estimation of a monotone dose–response curve. In particular, we establish novel theory and methods that have important practical implications. First, we provide conditions under which the estimator converges in distribution even when one of the nuisance estimators that are involved in the problem is inconsistent. This contrasts with the results in Westling and Carone (2019), which required that both nuisance parameters be estimated consistently. We also propose two estimators of the scale parameter arising in the limit distribution, one of which requires both nuisance estimators to be consistent, and the other of which does not. Second, we demonstrate that our estimator is invariant to strictly monotone transformations of the exposure. Third, we study the joint convergence of our proposed estimator at two points and use this result to construct confidence intervals for causal effects. Fourth, we study the behaviour of our estimator in the context of discrete exposures. Fifth, we propose an alternative estimator based on cross-fitting the nuisance estimators and demonstrate that this strategy removes the need for empirical process conditions that were required in Westling and Carone (2019). Finally, we investigate the behaviour of our estimator in comprehensive numerical studies and compare its behaviour with that of the local linear estimator of Kennedy et al. (2017).
The remainder of the paper is organized as follows. In Section 2, we concretely define the estimator proposed. In Section 3, we study theoretical properties of the estimator. In Section 4, we propose methods for pointwise inference. In Section 5, we perform numerical studies to assess the performance of the estimator and, in Section 6, we use this procedure to investigate the relationship between BMI and immune response to human immunodeficiency virus (HIV) vaccines by using data from several randomized trials. Finally, we provide concluding remarks in Section 7. Proofs of all theorems are provided in on-line supplementary material.
Some example data and the programs that were used to analyse them can be obtained from
https://dbpia.nl.go.kr/jrsssb/issue/.
2 Proposed approach
2.1 Review of isotonic regression
The isotonic regression estimator rn has many attractive properties. First, unlike smoothing-based estimators, isotonic regression does not require the choice of a kernel function, bandwidth or any other tuning parameter. Second, it is invariant to strictly increasing transformations of A. Specifically, if is a strictly increasing function, and is the isotonic regression of Y1, Y2, …, Yn on H(A1), H(A2), …, H(An), it follows that . Third, rn is uniformly consistent on any strict subinterval of . Fourth, n1/3{rn(a) − r0(a)} converges in distribution to for any interior point a of at which , and exist and are positive and continuous in a neighbourhood of a. Here, , where Z0 denotes a two-sided Brownian motion originating from zero, and is said to follow Chernoff’s distribution. Chernoff’s distribution has been extensively studied: among other properties, it is a log-concave and symmetric law centred at zero, has moments of all orders and can be approximated by an N(0, 0.52) distribution (Chernoff, 1964; Groeneboom and Wellner, 2001). It appears often in the limit distribution of monotonicity-constrained estimators.
2.2 Definition of proposed estimator
In our setting, is known to be monotone. Therefore, instead of using a local linear regression to estimate the conditional mean of the pseudo-outcomes, it is natural to consider as an estimator the isotonic regression of the pseudo-outcomes on A1, A2, …, An. Using the GCM representation of isotonic regression that was stated in the previous section, we can summarize our estimation procedure as follows.
Step 1: construct estimators and gn of μ0 and g0 respectively.
- Step 2: for each a in the unique values of A1, A2, …, An, compute and set(1)
Step 3: compute the GCM of the set of points {(0, 0)}∪{(Fn(Ai), Γn(Ai)):i = 1, 2, …, n} over [0, 1].
Step 4: define θn(a) as the left derivative of evaluated at Fn(a).
If were only known to be monotone on a fixed subinterval , we would define as the marginal distribution function restricted to , and Fn as its empirical counterpart. Similarly, in expression (1) would be replaced with . In all other respects, our estimation procedure would remain the same.
3 Theoretical properties
3.1 Invariance to strictly increasing exposure transformations
An important feature of the estimator proposed is that, as with the isotonic regression estimator, it is invariant to any strictly increasing transformation of A. This is a desirable property because the scale of a continuous exposure is often arbitrary from a statistical perspective. For instance, if A is temperature, whether A is measured in degrees Fahrenheit or Celsius or in kelvins does not change the information that is available. In particular, if the parameters and correspond to using as exposure A and H(A) respectively for H some strictly increasing transformation, then and encode exactly the same information about the effect of A on Y after adjusting for W. It is therefore natural to expect any sensible estimator to be invariant to the scale on which the exposure is measured.
We note that it is easy to ensure that and . Set U: = Fn(A), which is also equal to , and let be an estimator of the conditional mean of Y given (U, W) = (u, w). Then, taking , we have that satisfies the desired property. Similarly, letting be an estimator of the conditional density of U = u given W = w and, setting , we may take .
3.2 Consistency
We now provide sufficient conditions under which consistency of θn is guaranteed. Our conditions require controlling the uniform entropy of certain classes of functions. For a uniformly bounded class of functions , a finite discrete probability measure Q, and any ɛ > 0, the ɛ-covering-number of relative to the L2(Q) metric is the smallest number of L2(Q) balls of radius less than or equal to ɛ needed to cover . The uniform ɛ-entropy of is then defined as , where the supremum is taken over all finite discrete probability measures. For a thorough treatment of covering numbers and their role in empirical process theory, we refer readers to van der Vaart and Wellner (1996).
Below, we state three sufficient conditions that we shall refer to in the following theorem.
Condition 1There exist constants and V ∈ [0, 2) such that, almost surely as n → ∞, and gn are contained in classes of functions and respectively, satisfying
- (a)
for all , and K1 ⩽ g ⩽ K2 for all , and
- (b)
and for all ɛ ⩽ δ.
Condition 2There exist and such that and .
Condition 3There exist subsets and of such that and
- (a)
for all ,
- (b)
for all and
- (c)
and for all .
Under these three conditions, we have the following result.
Theorem 1(consistency)
If conditions 1–3 hold, then for any value such that , is continuous at a and F0 is strictly increasing in a neighbourhood of a. If is uniformly continuous and F0 is strictly increasing on , then for any bounded strict subinterval .
We note that, in the pointwise statement of theorem 1, F0(a) is required to be in the interior of [0, 1] and, similarly, the uniform statement of theorem 1 covers only strict subintervals of . This is due to the well-known boundary issues with Grenander-type estimators. Various remedies have been proposed in particular settings, and it would be interesting to consider these in future work (see, for example, Woodroofe and Sun (1993), Balabdaoui et al. (2011) and Kulikov and Lopuhaä (2006)).
Condition 1 requires that and gn eventually be contained in uniformly bounded function classes that are sufficiently small for certain empirical process terms to be controlled. This condition is easily satisfied if, for instance, and are parametric classes. It is also satisfied for many infinite dimensional function classes. Uniform entropy bounds for many such classes may be found in chapter 2.6 of van der Vaart and Wellner (1996). We note that there is an asymmetry between the entropy requirements for and in part (b) of condition 1. This is due to the term appearing in Γn(a). To control this term, we use an upper bound of the form from the theory of empirical U-processes (Nolan and Pollard, 1987)—this contrasts with the uniform entropy integral that bounds ordinary empirical processes indexed by a uniformly bounded class . In Section 3.7, we consider the use of cross-fitting to avoid the entropy conditions in condition 1.
Condition 2 requires that and gn tend to limit functions and , and condition 3 requires that either or for (F0 × Q0) almost every (a, w). If either
- (a)
and are null sets or
- (b)
and are null sets,
then condition 3 is known simply as double robustness of the estimator θn relative to the nuisance functions μ0 and g0: θn is consistent as long as or . Doubly robust estimators are at this point a mainstay of causal inference and have been studied for over two decades (see, for example, Robins et al. (1994), Rotnitzky et al. (1998), Scharfstein et al. (1999), van der Laan and Robins (2003), Neugebauer and van der Laan (2005) and Bang and Robins (2005)). However, condition 3 is more general than classical double robustness, as it allows neither nor gn to tend to their true counterparts over the whole domain, as long as at least one of or gn tends to the truth for almost every point in the domain.
3.3 Convergence in distribution
Condition 4There exists ɛ0 > 0 such that
- (a)
,
- (b)
and
- (c)
.
Condition 5and are continuously differentiable in a neighbourhood of a uniformly over .
Under conditions that have been introduced so far, we have the following distributional result.
Theorem 2(convergence in distribution)
If conditions 1–5 hold, thenfor any such that , where follows the standard Chernoff distribution andwith denoting .
We note that the limit distribution in theorem 2 is the same as that of the standard isotonic regression estimator up to a scale factor. As noted above, when either
- (a)
Y and W are independent given A or
- (b)
A is independent of W,
the functions and r0 coincide. As such, we can directly compare the respective limit distributions of and n1/3{rn(a) − r0(a)} under these conditions. When both and , rn(a) is asymptotically more concentrated than θn(a) in scenario (a), and less concentrated in scenario (b). This is analogous to findings in linear regression, where including a covariate that is uncorrelated with the outcome inflates the standard error of the estimator of the coefficient corresponding to the exposure, whereas including a covariate that is correlated with the outcome but uncorrelated with the exposure deflates its standard error.
Condition 4 requires that, on the set where is consistent but gn is not, converges faster than n−1/3 uniformly in a neighbourhood of a, and similarly for gn on the set . On the set where both and gn are consistent, only the product of their rates of convergence must be faster than n−1/3. Hence, a non-degenerate limit theory is available as long as at least one of the nuisance estimators is consistent at a rate faster than n−1/3, even if the other nuisance estimator is inconsistent. This suggests the possibility of performing doubly robust inference for , i.e. of constructing confidence intervals and tests based on θn(a) with valid calibration even when one of μ0 and g0 is inconsistently estimated. This is explored in Section 4. Finally, as in theorem 1, we allow that neither nor gn be consistent everywhere, as long as for (F0 × Q0) almost every (a, w) at least one of or gn is consistent.
We remark that, if it is known that is consistent for μ0(a, ·) in an L2(Q0) sense at rate faster than n−1/3, the isotonic regression of the plug-in estimator —which can be equivalently obtained by setting in the construction of θn(a)—achieves a faster rate of convergence to than does θn(a). This might motivate an analyst to use rather than θn(a) in such a scenario. However, the consistency of hinges entirely on the fact that and, in particular, will be inconsistent if , even if . Additionally, the estimator may not generally admit a tractable limit theory on which to base the construction of valid confidence intervals, particularly when machine learning methods are used to build .
3.4 Grenander-type estimation without domain transformation
3.5 Discrete domains
In some circumstances, the exposure A is discrete rather than continuous. Our estimator works equally well in these cases, since, as we highlight below, it turns out then to be asymptotically equivalent to the well-studied AIPW estimator. As a result, the large sample properties of our estimator can be derived from the large sample properties of the AIPW estimator, and asymptotically valid inference can be obtained by using standard influence-function-based techniques.
The large sample properties of , including doubly robust consistency and convergence in distribution at the regular parametric rate n−1/2, are well established (Robins et al., 1994). Therefore, many properties of θn in this case can be determined by using the results of Westling et al. (2018), who studied the behaviour of the isotonic correction of an initial estimator. In particular, as long as is non-decreasing on . Uniform consistency of over thus implies uniform consistency of θn. Furthermore, if is strictly increasing on and converges in distribution, then , so large sample standard errors for are also valid for θn. If is not strictly increasing on but instead has flat regions, then θn is more efficient than on these regions, and confidence intervals centred near θn but based on the limit theory for will be conservative.
3.6 Large sample results for causal effects
In many applications, in addition to the causal dose–response curve a↦m0(a) itself, causal effects of the form (a1, a2)↦m0(a1) − m0(a2) are of scientific interest as well. Under the identification conditions that were discussed in Section 1.2 applied to each of a1 and a2, such causal effects are identified with the observed data parameter . A natural estimator for such a causal effect in our setting is θn(a1) − θn(a2). If the conditions of theorem 1 hold for both a1 and a2, then the continuous mapping theorem implies that . However, since theorem 2 provides only marginal distributional results, and thus does not describe the joint convergence of , it cannot be used to determine the large sample behaviour of . The following result demonstrates that such joint convergence can be expected under the aforementioned conditions, and that the bivariate limit distribution of Zn(a1, a2) has independent components.
Theorem 3(joint convergence in distribution)
If conditions 1–5 hold for and , then Zn(a1, a2) converges in distribution to , where and are independent standard Chernoff distributions and the scale parameter τ0 is as defined in theorem 2.
Theorem 3 implies that, under the conditions stated, converges in distribution to .
3.7 Use of cross-fitting to avoid empirical process conditions
Theorems 1 and 2 reveal that the statistical properties of θn depend on the nuisance estimators and gn in two important ways. First, we require in condition 1 that or gn fall in sufficiently small classes of functions, as measured by metric entropy, to control certain empirical process remainder terms. Second, we require in conditions 2 and 3 that at least one of or gn be consistent almost everywhere (for consistency), and in condition 4 that the product of their rates of convergence be faster than n−1/3 (for convergence in distribution). In observational studies, researchers can rarely specify a priori correct parametric models for μ0 and g0. This motivates use of data-adaptive estimators of these nuisance functions to meet the second requirement. However, such estimators often lead to violations of the first requirement, or it may be onerous to determine that they do not. Thus, because it may be difficult to find nuisance estimators that are both sufficiently data adaptive to meet required rates of convergence and fall in sufficiently small function classes to make empirical process terms negligible, simultaneously satisfying these two requirements can be challenging in practice.
As we now demonstrate, utilizing the cross-fitted estimator enables us to avoid the empirical process condition 1. We first introduce the following two conditions, which are analogues of conditions 1 and 2.
Condition 6There exist constants such that, almost surely as n → ∞ and for all v, μn, v and gn, v are contained in classes of functions and respectively, satisfying
- (a)
for all , and for all , and
- (b)
for almost all a and w.
Condition 7There exist and such that and .
We then have the following analogue of theorem 1 establishing consistency of the cross-fitted estimator .
Theorem 4(consistency of the cross-fitted estimator)
If conditions 6 and 7 and 3 hold, then for any such that , is continuous at a, and F0 is strictly increasing in a neighbourhood of a. If is uniformly continuous and F0 is strictly increasing on , then for any bounded strict subinterval .
For convergence in distribution, we introduce the following analogue of condition 4.
Condition 8There exists ɛ0 > 0 such that
- (a)
,
- (b)
and
- (c)
.
We then have the following analogue of theorem 2 for the cross-fitted estimator .
Theorem 5(convergence in distribution for the cross-fitted estimator)
If conditions 6, 7, 3, 8 and 5 hold, then for any such that .
The conditions of theorems 4 and 5 are analogous to those of theorems 1 and 2, with the important exception that the entropy condition 1(b) is no longer required. Therefore, the estimators μn, v and gn, v may be as data adaptive as one desires without concern for empirical process terms, as long as they satisfy the boundedness conditions that are stated in condition 6.
4 Construction of confidence intervals
4.1 Wald-type confidence intervals
We note that with . This suggests that we could either estimate and f0 separately and consider the ratio of these estimators, or that we could instead estimate directly and compose it with the estimator of F0 that is already available. The latter approach has the desirable property that the resulting scale estimator is invariant to strictly monotone transformations of the exposure. As such, this is the strategy that we favour. To estimate , we recall that the estimator ψn from Section 2 is a step function and is therefore not differentiable. A natural solution consists of computing the derivative of a smoothed version of ψn. We have found local quadratic kernel smoothing of points {(uj, ψn(uj)):j = 1, 2, …, K}, for uj the midpoints of the jump points of ψn, to work well in practice.
Theorem 3 can be used to construct Wald-type confidence intervals for causal effects of the form . We first construct estimates τn(a1) and τn(a2) of the scale parameters τ0(a1) and τ0(a2) respectively, and then compute an approximation of the (1 − α/2)-quantile of , where and are independent Chernoff distributions, using Monte Carlo simulations, for example. An asymptotic (1 − α)-level Wald-type confidence interval for is then .
In the next two subsections, we discuss different strategies for estimating the scale factor .
4.2 Scale estimation relying on consistent nuisance estimation
4.3 Doubly robust scale estimation
As noted above, theorem 2 provides the limit distribution of even if one of the nuisance estimators is inconsistent, as long as the consistent nuisance estimator converges sufficiently fast. We now show how we may capitalize on this result to provide a doubly robust estimator of . Since ψn is itself a doubly robust estimator of ψ0, so will be the proposed estimator of and hence also of the resulting estimator τn(a) of τ0(a). This contrasts with the estimator of that was described in the previous section, which required the consistency of both and gn.
To determine an appropriate value of the bandwidth h in practice, we propose the following empirical criterion. We first define the integrated scale and construct the estimator for any candidate h > 0. Furthermore, we observe that , which suggests the use of the empirical estimator . This motivates us to define , i.e. the value of h that makes γn(h) and closest. The proposed doubly robust estimator of is thus .
We make two final remarks regarding this doubly robust estimator of . First, we note that this estimator depends only on A and a through the ranks Fn(A) and Fn(a). Hence, as before, our estimator is invariant to strictly monotone transformations of the exposure A. Second, we note that, if does not depend on w and gn = 1, κn,DR(a) tends to the conditional variance , which is precisely the scale parameter appearing in standard isotonic regression.
4.4 Confidence intervals via sample splitting
5 Numerical studies
In this section, we perform numerical experiments to assess the performance of the proposed estimators of and of the three approaches for constructing confidence intervals, which we also compare with that of the local linear estimator and associated confidence intervals that were proposed in Kennedy et al. (2017).
In our experiments, we simulate data as follows. First, we generate as a vector of four independent standard normal variates. A natural next step would be to generate A given W. However, since our estimation procedures requires estimating the conditional density of U: = F0(A) given W, we instead generate U given W, and then transform U to obtain A. This strategy makes it easier to construct correctly specified parametric nuisance estimators in the context of these simulations. Given W = w, we generate U from the distribution with conditional density function for . We note that for all u ∈ [0, 1] and , and, also, that , so that U is marginally uniform. We then take A to be the evaluation at U of the quantile function of an equal weight mixture of two normal distributions with means −2 and 2 and standard deviation 1, which implies that A is marginally distributed according to this bimodal normal mixture. Finally, conditionally on A = a and W = w, we simulate Y as a Bernoulli random variate with conditional mean function given by , where denotes (1, w). We set , γ1 = (−1, −1, −1, 1, 1)T, γ2 = (3, −1, −1, 1, 1)T and γ3 = 3 in the experiments that we report on.
We estimate the true confounder-adjusted dose–response curve by using the causal isotonic regression estimator θn, the local linear estimator of Kennedy et al. (2017) and the sample splitting version of θn that was proposed by Banerjee et al. (2019) with m = 5 splits. For the local linear estimator, we use the data-driven bandwidth selection procedure that was proposed in section 3.5 of Kennedy et al. (2017). We consider three settings in which either both and gn are consistent, only is consistent or only gn consistent. To construct a consistent estimator , we use a correctly specified logistic regression model, whereas, to construct a consistent estimator gn, we use a maximum likelihood estimator based on a correctly specified parametric model. To construct an inconsistent estimator , we still use a logistic regression model but omit covariates W3 and W4 and all interactions. To construct an inconsistent estimator gn, we posit the same parametric model as before but omit W3 and W4. We construct pointwise confidence intervals for in each setting, using the Wald-type construction described in Section 4 with both the plug-in and the doubly robust estimators of . We expect intervals based on the doubly robust estimator of to provide asymptotically correct coverage rates for for each of the three settings, but we expect only asymptotically correct coverage rates in the first setting when the plug-in estimator of is used. We construct pointwise confidence intervals for the local linear estimator by using the procedure that was proposed in Kennedy et al. (2017), and for the sample splitting procedure by using the procedure that was discussed in Section 4.4. We consider the performance of these inferential procedures for values of a between −3 and 3.
Fig. 1(a) shows a single sample path of the causal isotonic regression estimator based on a sample of size n = 5000 and consistent estimators and gn. Also included in Fig. 1(a) are asymptotic 95% pointwise confidence intervals constructed by using the doubly robust estimator of . Fig. 1(b) shows the unadjusted isotonic regression estimate based on the same data and corresponding 95% asymptotic confidence intervals. The true causal and unadjusted regression curves are the smooth curves. We note that for a ≠ 0, since the relationship between Y and A is confounded by W, and indeed the unadjusted regression curve does not have a causal interpretation. Therefore, the marginal isotonic regression estimator will not be consistent for the true causal parameter. In this data-generating setting, the causal effect of A on Y is larger than the marginal effect of A on Y in the sense that has greater variation over values of a than does r0(a).

(a) Causal isotonic regression estimate using consistent nuisance estimators and gn, and (b) regular isotonic regression estimate: , pointwise 95% confidence intervals constructed by using the doubly robust estimator;
, true functions
We perform 1000 simulations, each with n ∈ {500, 1000, 2500, 5000} observations. Fig. 2 displays the empirical standard error of the three estimators considered over these 1000 simulated data sets as a function of a and for each value of n. We first note that the standard error of the local linear estimator is smaller than that of θn, which is expected because of the faster rate of convergence of the local linear estimator. The sample splitting procedure also reduces the standard error of θn. Furthermore, the standard deviation of the local linear estimator appears to decrease faster than n−1/3, whereas the standard deviation of the estimators based on θn do not, in line with the theoretical rates of convergence of these estimators. We also note that inconsistent estimation of the propensity has little effect on the standard errors of any of the estimators, but inconsistent estimation of the outcome regression results in slightly larger standard errors.

Standard error of the three estimators scaled by n1/3 as a function of n for various values of a and in contexts in which and gn are either consistent or inconsistent, computed empirically over 1000 simulated data sets of different sizes (, sample size 500;
, sample size 1000;
, sample size 2500;
, sample size 5000): (a) correct outcome regression, correct propensity; (b) correct outcome regression, incorrect propensity; (c) incorrect outcome regression, correct propensity
Fig. 3 displays the absolute bias of the three estimators. For most values of a, the estimator θn that is proposed here has smaller absolute bias than the local linear estimator, and its absolute bias decreases faster than n−1/3. The absolute bias of the local linear estimator depends strongly on a, and in particular is largest where the second derivative of is large in absolute value, agreeing with the large sample theory that was described in Kennedy et al. (2017). The sample splitting estimator has larger absolute bias than does θn because it inherits the bias of θn/m. The bias is especially large for values of a in the tails of the marginal distribution of A.

Absolute bias of the three estimators scaled by n1/3 as a function of n for various values of a and in contexts in which and gn are either consistent or inconsistent, computed empirically over 1000 simulated data sets of different sizes (, sample size 500;
, sample size 1000;
, sample size 2500;
, sample size 5000): (a) correct outcome regression, correct propensity; (b) correct outcome regression, incorrect propensity; (c) incorrect outcome regression, correction propensity
Fig. 4 shows the empirical coverage of nominal 95% pointwise confidence intervals for a range of values of a for the four methods that were considered. For both the plug-in and the doubly robust intervals centred near θn, the coverage improves as n grows, especially for values of a in the tails of the marginal distribution of A. Under correct specification of outcome and propensity regression models, the plug-in method attains close to nominal coverage rates for a between −3 and 3 by n = 1000. When the propensity estimator is inconsistent, the plug-in method still performs well in this example, although we do not expect this to be the case always. However, when is inconsistent, the plug-in method is very conservative for positive values of a. The doubly robust method attains close to nominal coverage for large samples as long as one of gn or is consistent. Compared with the plug-in method, the doubly robust method requires larger sample sizes to achieve good coverage, especially for extreme values of a. This is because the doubly robust estimator of has a slower rate of convergence than does the plug-in estimator, as demonstrated by box plots of these estimators that are provided in the on-line supplementary material.

Observed coverage of pointwise 95% confidence intervals by using θn and the plug-in method (top row), θn and the doubly robust method (second row), the local linear estimator and associated intervals (third row) and the sample splitting estimator (bottom row), considered for various values of a and computed empirically over 1000 simulated data sets of different sizes (, nominal coverage rate;
, sample size 500;
sample size 1000;
, sample size 2500;
, sample size 5000): (a) correct outcome regression, correct propensity; (b) correct outcome regression, incorrect propensity; (c) incorrect outcome regression, correct propensity
The confidence intervals that are associated with the local linear estimator have poor coverage for values of a where the bias of the estimator is large, which, as mentioned above, occurs when the second derivative of is large in absolute value. Overall, the sample splitting estimator has excellent coverage, except perhaps for values of a in the tails of the marginal distribution of A when n is small or moderate, in which case the coverage is near 90%.
We also conducted a small simulation study to illustrate the performance of the proposed procedures when machine learning techniques are used to construct and gn. To estimate μ0 consistently, we used Super Learner (van der Laan et al., 2007) with a library consisting of generalized linear models, multivariate adaptive regression splines and generalized additive models. To estimate g0 consistently, we used the method that was proposed by Díaz and van der Laan (2011) with covariate vector (W1, W2, W3, W4). To produce inconsistent estimators or gn, we used the same estimators but omitted covariates W1 and W2. We also considered the estimator that was obtained via cross-fitting these nuisance parameters, as discussed in Section 3.7, as well as the local linear estimator. Because of computational limitations, we performed 1000 simulations at sample size n = 1000 only. Fig. 5 shows the coverage of nominal 95% confidence intervals. The plug-in intervals achieve very close to nominal coverage under consistent estimation of both nuisances, and they also achieve surprisingly good coverage rates when the propensity is inconsistently estimated. The plug-in intervals are somewhat conservative when the outcome regression is inconsistently estimated. The doubly robust method is anti-conservative under inconsistent estimation of both nuisances and also when the propensity is inconsistently estimated, with coverage rates mostly between 90% and 95%. Good coverage rates are also achieved when the outcome regression is inconsistently estimated. These results suggest that the doubly robust intervals may require larger sample sizes to achieve good coverage, particularly when machine learning estimators are used for and gn. The plug-in intervals appear to be relatively robust to moderate misspecification of models for the nuisance parameters in smaller samples. Histograms of the estimators of and are provided in the on-line supplementary material. Confidence intervals based on the local linear estimator show a similar pattern to that in the previous simulation study, undercovering where the second derivative of the true function is large in absolute value. Cross-fitting had little effect on coverage.

Observed coverage of pointwise 95% doubly robust and plug-in confidence intervals by using machine learning estimators based on simulated data including n = 1000 observations (, nominal coverage rate;
, causal isotonic plus plug in;
, causal isotonic plus doubly robust;
, cross-fitted causal isotonic plus plug in;
, cross-fitted causal isotonic plus doubly robust;
, local linear): (a) correct outcome regression, correct propensity; (b) correct outcome regression, incorrect propensity; (c) incorrect outcome regression, correct propensity
As noted above, we found in our numerical experiments that the plug-in estimator of the scale parameter was surprisingly robust to inconsistent estimation of the nuisance parameters, whereas its doubly robust estimator was anticonservative even when the nuisance parameters were estimated consistently. This can be explained in terms of the bias and variance of the two scale estimators proposed. On one hand, under inconsistent estimation of any nuisance function, the plug-in estimator of the scale parameter is biased, even in large samples. However, its variance decreases relatively quickly with sample size, since it is a simple empirical average of estimated functions. On the other hand, the doubly robust estimator is asymptotically unbiased, but its variance decreases much slower with sample size. These trends can be observed in the figures that are provided in the on-line supplementary material. In sufficiently large samples, the doubly robust estimator is expected to outperform the plug-in estimator in terms of mean-squared error when one of the nuisances is inconsistently estimated. However, the sample size that is required for this trade-off to affect confidence interval coverage significantly depends on the degree of inconsistency. Although we did not see this trade-off at the sample sizes that were used in our numerical experiments, we expect the benefits of the doubly robust confidence interval construction to become apparent in smaller samples in other settings.
6 Body mass index and T-cell response in human immunodeficiency virus vaccine studies
The scientific literature indicates that, for several vaccines, obesity or BMI is inversely associated with immune responses to vaccination (see, for example, Sheridan et al. (2012), Young et al. (2013), Jin et al. (2015), Painter et al. (2015) and Liu et al. (2017)). Some of this literature has investigated potential mechanisms of how obesity or higher BMI might lead to impaired immune responses. For example, Painter et al. (2015) concluded that obesity may alter cellular immune responses, especially in adipose tissue, which varies with BMI. Sheridan et al. (2012) found that obesity is associated with decreased CD8+ T-cell activation and decreased expression of functional proteins in the context of influenza vaccines. Liu et al. (2017) found that obesity reduced hepatitis B immune responses through ‘leptin-induced systemic and B cell intrinsic inflammation, impaired T cell responses and lymphocyte division and proliferation’. Given this evidence of a monotone effect of BMI on immune responses, we used the methods that are presented in this paper to assess the covariate-adjusted relationship between BMI and CD4+ T-cell responses using data from a collection of clinical trials of candidate HIV vaccines. We present the results of our analyses here.
Jin et al. (2015) compared the rate of CD4+ T cell response with HIV peptide pools among low (BMI <25) medium (25 ⩽ BMI <30) and high (BMI ⩾ 30) BMI participants, and they found that low BMI participants had a statistically significantly greater response rate than high BMI participants by using Fisher’s exact test. However, such a marginal assessment of the relationship between BMI and immune response can be misleading because there are known common causes, such as age and sex, of both BMI and immune response. For this reason, Jin et al. (2015) also performed a logistic regression of the binary CD4+ responses against sex, age, BMI (not discretized), vaccination dose and number of vaccinations. In this adjusted analysis, they found a significant association between BMI and CD4+ cell response rate after adjusting for all other covariates (odd ratio 0.92; 95% confidence interval 0.86, 0.98; p = 0.007). However, such an adjusted odds ratio has a formal causal interpretation only under strong parametric assumptions. As discussed in Section 1.2, the covariate-adjusted dose–response function is identified with the causal dose–response curve without making parametric assumptions and is therefore of interest for understanding the continuous covariate-adjusted relationship between BMI and immune responses.
We note that there is some debate in the causal inference literature about whether exposures such as BMI have a meaningful interpretation in formal causal modelling. In particular, some researchers suggest that causal models should always be tied to hypothetical randomized experiments (see, for example Bind and Rubin (2019)), and it is difficult to imagine a hypothetical randomized experiment that would assign participants to levels of BMI. From this perspective, it may therefore not be sensible to interpret in a causal manner in the context of this example. Nevertheless, as discussed in Section 1, we contend that is still of interest. In particular, it provides a meaningful summary of the relationship between BMI and immune response accounting for measured potential confounders. In this case, we interpret as the probability of immune response in a population of participants with BMI-value a but sex, age, vaccination dose, number of vaccinations and study with a similar distribution to that of the entire study population.
We pooled data from the vaccine arms of 11 phase I–II clinical trials, all conducted through the HIV Vaccine Trials Network. 10 of these trials were previously studied in the analysis that was presented in Jin et al. (2015), and a detailed description of the trials are contained therein. The final trial in our pooled analysis is HVTN 100, in which 210 participants were randomized to receive four doses of the ALVAC-HIV vaccine (vCP1521). The ALVAC-HIV vaccine, in combination with an AIDSVAX boost, was found to have statistically significant vaccine efficacy against HIV type 1 in the RV-144 trial that was conducted in Thailand (Rerks-Ngarm et al., 2009). CD4+ and CD8+ T-cell responses to HIV peptide pools were measured in all 11 trials by using validated intracellular cytokine staining at HIV Vaccine Trials Network laboratories. These continuous responses were converted to binary indicators of whether there was a significant change from baseline by using the method described in Jin et al. (2015). We analysed these binary responses at the first visit following administration of the last vaccine dose—either 2 or 4 weeks after the final vaccination depending on the trial. After accounting for missing responses from a small number of participants, our analysis data sets consisted of a total of n = 439 participants for the analysis of CD4+ cell responses and n = 462 participants for CD8+ cell responses. Here, we focus on analysing CD4+ cell responses; we present the analysis of CD8+ cell responses in the on-line supplementary material.
We assessed the relationship between BMI and T-cell response by estimating the covariate-adjusted dose–response function using our cross-fitted estimator , the local linear estimator and the sample splitting version of our estimator with m = 5 splits. We adjusted for sex, age, vaccination dose, number of vaccinations and study. We estimated μ0 and g0 as in the machine-learning-based simulation study that was described in Section 5, and we constructed confidence intervals for our estimator by using both the plug-in and the doubly robust estimators that were described above.
Fig. 6 presents the estimated probability of a positive CD4+ T-cell response as a function of BMI for BMI-values between the 0.05- and 0.95-quantile of the marginal empirical distribution of BMI by using our estimator (Fig. 6(a)), the local linear estimator (Fig. 6(b)) and the sample-splitting estimator (Fig. 6(c)). Pointwise 95% confidence intervals are shown as broken or dotted curves. The three methods found qualitatively similar results. We found that the change in probability of CD4+ cell response appears to be largest for BMI <20 and BMI >30. We estimated the probability of having a positive CD4+ T-cell response, after adjusting for potential confounders, to be 0.52 (95% doubly robust confidence interval 0.44–0.59) for a BMI of 20, 0.47 (0.42–0.52) for a BMI of 25, 0.47 (0.32–0.62) for a BMI of 30 and 0.29 (0.12–0.47) for a BMI of 35. We estimated the difference between these probabilities for BMIs of 20 and 35 to be 0.22 (confidence interval 0.03–0.41).

Estimated probabilities of CD4+ T-cell response and 95% pointwise confidence intervals as a function of BMI, adjusted for sex, age, number of vaccinations received, vaccine dose and study: (a) estimator proposed here; , confidence intervals based on the plug-in estimator of the scale parameter;
, confidence interval based on the doubly robust estimator of the scale parameter; (b) local linear estimator of Kennedy et al. (2017); (c) sample splitting version of our estimator with m = 5 splits
7 Concluding remarks
The work that we have presented in this paper lies at the interface of causal inference and shape-constrained non-parametric inference, and there are natural future directions building on developments in either of these areas. Inference on a monotone causal dose–response curve when outcome data are only observed subject to potential coarsening, such as censoring, truncation or missingness, is needed to increase the applicability of our proposed method. To tackle such cases, it appears most fruitful to follow the general primitive strategy that was described in Westling and Carone (2019) based on a revised causal identification formula allowing such coarsening.
It would be useful to develop tests of the monotonicity assumption, like Durot (2003) did for regression functions. Such a test could probably be developed by studying the large sample behaviour of under the null hypothesis that is monotone, where Ψn and are the primitive estimator and its greatest convex minorant as defined in Section 2.2. Such a result would probably permit testing with a given asymptotic size when is strictly increasing, and asymptotically conservative inference otherwise. It would also be useful to develop methods for uniform inference. Uniform inference is difficult in this setting because does not convergence weakly as a process in the space of bounded functions on to a tight limit process. Indeed, theorem 3 indicates that converges to an independent white noise process, which is not tight, so this convergence is not useful for constructing uniform confidence bands. Instead, it may be possible to extend the work of Durot et al. (2012) to our setting (and other generalized Grenander-type estimators) by demonstrating that converges in distribution to a non-degenerate limit for some constant depending on , a deterministic sequence cn and a suitable sequence of subsets increasing to . Developing procedures for uniform inference and tests of the monotonicity assumption are important areas for future research.
An alternative approach to estimating a causal dose–response curve is to use local linear regression, like Kennedy et al. (2017) did. As is true in the context of estimating classical univariate functions such as density, hazard and regression functions, there are certain trade-offs between local linear smoothing and monotonicity-based methods. On the one hand, local linear regression estimators exhibit a faster n−2/5-rate of convergence whenever optimal tuning rates are used and the true function has two continuous derivatives. However, the limit distribution involves an asymptotic bias term depending on the second derivative of the true function, so confidence intervals based on optimally chosen tuning parameters provide asymptotically correct coverage only for a smoothed parameter rather than the true parameter of interest. In contrast, monotonicity-constrained estimators such as the estimator that is proposed here exhibit a rate of convergence of n−1/3 whenever the true function is strictly monotone and has one continuous derivative, do not require choosing a tuning parameter, are invariant to strictly increasing transformations of the exposure and their limit theory does not include any asymptotic bias (as illustrated by theorem 2). We note that both estimators achieve the optimal rate of convergence for pointwise estimation of a univariate function under their respective smoothness constraints. In our view, the ability to perform asymptotically valid inference by using a monotonicity-constrained estimator is one of the most important benefits of leveraging the monotonicity assumption rather than using smoothing methods. This advantage was evident in our numerical studies when comparing the isotonic estimator that is proposed here and the local linear method of Kennedy et al. (2017). Undersmoothing can be used to construct calibrated confidence intervals by using kernel smoothing estimators, but performing adequate undersmoothing in practice is challenging.
The two methods for pointwise asymptotic inference that we presented require estimation of the derivative and the scale parameter . We found that the plug-in estimator of had low variance but possibly large bias depending on the levels of inconsistency of and gn, and that its doubly robust estimator instead had high variance but low bias as long as either or gn is consistent. In practice, we found that the low variance of the plug-in estimator often outweighed its bias, resulting in better coverage rates for intervals based on the plug-in estimator of , especially in samples of small and moderate sizes. Whether a doubly robust estimator of with smaller variance can be constructed is an important question to be addressed in future work. We found that sample splitting with as few as m = 5 splits provided doubly robust coverage, and the sample splitting estimator also had smaller variance than the original estimator, at the expense of some additional bias.
It would be even more desirable to have inferential methods that do not require estimation of additional nuisance parameters or sample splitting. Unfortunately, the standard non-parametric bootstrap is not generally consistent in Grenander-type estimation settings and, although alternative bootstrap methods have been proposed, to our knowledge, all such proposals require the selection of critical tuning parameters (Kosorok, 2008; Sen et al., 2010). Likelihood-ratio-based inference for Grenander-type estimators has proven fruitful in a variety of contexts (see, for example, Banerjee and Wellner (2001) and Groeneboom and Jongbloed (2015)), and extending such methods to our context is also an area of significant interest in future work.
Supporting information
Additional ‘supporting information’ may be found in the on-line version of this article:
‘Supplementary material for: Causal isotonic regression’.
Acknowledgements
Research reported in this publication was supported by the National Institute of Allergy and Infectious Diseases of the National Institutes of Health grant AI068635 (PBG) and by the National Lung, Heart, and Blood Institute of the National Institutes of Health grant R0IHL137808 (MC). The content is solely the responsibility of the authors and does not necessarily represent the official views of the National Institutes of Health. The authors are also grateful to the members of the University of Pennsylvania Center for Causal Inference working group for helpful feedback on an early version of this work, and to two reviewers and an Associate Editor for their detailed and insightful comments. The authors also thank the study participants and investigators of the HIV Vaccine Trials Network studies 044, 052, 060, 063, 064, 068, 069, 070, 080, 100 and 204.