886
Views
9
CrossRef citations to date
0
Altmetric
Theory and Methods

Active Clinical Trials for Personalized Medicine

Pages 875-887 | Received 01 Apr 2014, Published online: 18 Aug 2016
 

Abstract

Individualized treatment rules (ITRs) tailor treatments according to individual patient characteristics. They can significantly improve patient care and are thus becoming increasingly popular. The data collected during randomized clinical trials are often used to estimate the optimal ITRs. However, these trials are generally expensive to run, and, moreover, they are not designed to efficiently estimate ITRs. In this article, we propose a cost-effective estimation method from an active learning perspective. In particular, our method recruits only the “most informative” patients (in terms of learning the optimal ITRs) from an ongoing clinical trial. Simulation studies and real-data examples show that our active clinical trial method significantly improves on competing methods. We derive risk bounds and show that they support these observed empirical advantages. Supplementary materials for this article are available online.

View correction statement:
Correction

Appendix A: Technical Proofs

A.1. Intrinsic Dimension of supp(Π)

We explain the meaning of “intrinsic dimension” introduced in Assumption (A3) here. We say that supp(Π) possesses a tree decomposition T=Ti,j,i1,j=1...J(i) if

1.

T1, 1 = supp(Π), and {Ti, j}J(i)j = 1 forms a disjoint partition of supp(Π) for all i ⩾ 1.

2.

Nested partition: ∀i ⩾ 2, j = 1, …, J(i), there exists a unique 1 ⩽ kJ(i − 1) such that Ti, jTi − 1, k.

3.

Bounded diameter: for all i ⩾ 1, 1 ⩽ jJ(i), diam (Ti,j):=supx,yTi,jx-y2K12-ifor some K1 = K1(Π).

4.

Regularity: for any i ⩾ 1, 1 ⩽ jJ(i) and 0 < r ⩽ 2i the following holds: there exists a 1 ⩽ dp (d is the intrinsic dimension) such that for all xTi, j, (A.1) c1rdΠ(B(x,r)Ti,j)c2rd(A.1) for some 0 < c1(Π) ⩽ c2(Π) < ∞, which are independent of i, j. Here, B(x, r) is the Euclidean ball of radius r centered at x.

A simple example that gives a good intuition to the tree decomposition is the uniform distribution over the unit cube in Rp. In this case, the tree decomposition is given by partitioning the unit cube into dyadic cubes and d = p. If supp(Π) is contained in a proper subspace W of Rp, then ddim(W).

Let Bi be the sigma-algebra generated by the collection of sets {Ti, j, j = 1, …, J(i)} (a partition of supp(Π) on the level i). A regular approximation of ASk in Algorithm 1 is given by act k:=A:ABmk-1,A AS k.

A.2. Properties of Kernel Estimate

A.2.1. Preliminaries

For a measurable set S⊂supp(Π), define ΠS(dx) ≔ Π(dx|xS) as the conditional distribution on S, and set Qh(x|S):=RpKhx-ydΠS(y).Since Π is assumed to be known, we can directly compute Qh(x|S) now. Accordingly, we modify the original kernel estimate for ηj, that is, (Equation2.3), as follows: let {(X(i), A(i), R(i)), i = 1…N} be an iid sample from the conditional joint distribution of (X, A, R) given that XS, and set (A.2) η^j(x;h,S)=1Ni=1NR(i)I{A(i)=j}Khx-X(i)Qh(x|S)P(A(i)=j),j=±1,f^(x;h,S)=η^1(x;h,S)-η^-1(x;h,S).(A.2) We will discuss properties of these estimators in below.

Let h > 0, SBj, and h ⩽ 2j, and define (A.3) Qh,m(x|S):=Rpx-y2mKhx-ydΠS(y).(A.3) We next study the upper and lower bounds of Qh,m(x|S) based on Assumptions (A1)–(A4). Since K is bounded and compactly supported, there exists R = RK > 0 such that K(x) ⩽ ‖KI{xB(0, RK)}. Let F > 0 be a large enough constant, namely, Fd ⩾ 2c2/c1. Recall that c1, c2 are defined in (EquationA.1). Note that Assumption (A2) implies the following: (A.4) Qh,m(x|S)KB(x,h)Sx-y2mdΠS(y)Kh/FmB(x,h)SdΠS(y)-B(x,h/F)SdΠS(y)Kh/Fmc1hd-c2(h/F)dΠ(S)12FmKc1hd+mΠ(S):=c3hd+mΠ(S),(A.4) and (A.5) Qh,m(x|S)KB(x,RKh)Ax-y2mdΠS(y)KRKm+dc2×hd+mΠ(S):=c4hd+mΠ(S).(A.5)

In what follows, we will set Qh(x|S) ≔ Qh, 0(x) for brevity.

A.2.2. Some Bounds for the Kernel Estimators

In this subsection, we derive basic concentration inequalities for the kernel estimators of ηj(x)=E[R|A=j,X=x], j = ±1 restricted to S, that is, η^j(x;h,S) defined in (EquationA.2). The proof of the results can be found in the online supplementary materials.

Lemma A.1.

For all t > 0 satisfying t + d2log (1/h) ⩽ nhd, with probability ⩾ 1 − 2et, supx supp (Π)S|η^j(x;h)-ηj(x)|Ch+Π(S)(t+d2log(1/h))nhd,where C = C(M, c1, c2, L, LK, ‖K, ℓK, RK) is a constant.

The following corollary is immediate:

Corollary A.1.

Set hn ≔ {Π(S)(t + dlog (n/Π(S)))/n}1/(d + 2). Then, under assumptions of Lemma A.1, with probability ⩾ 1 − 4et, supx supp (Π)S|f^(x;hn)-f*(x)|4Chn,where constant C is the same as in Lemma A.1.

A.3. Proof of Theorem 3.1

A.3.1. Comparison Inequality

Our Lemma A.2 below illustrates the connection between the risk V(D^)-V(D*) of a treatment rule D^(x)=sign(f^(x)) and the sup-norm f^-f*, supp (Π).

Lemma A.2.

Under the margin assumption (A4), V(D^)-V(D*)C(γ)(f^-f*)Isign(f^)sign(f*), supp (Π)1+γ.

Proof.

It is easy to see that V(D^)-V(D*)=2E|f*(X)|I{D^(X)D*(X)}. The rest of the argument repeats lemma 5.1 in Audibert and Tsybakov (Citation2007).

A.3.2. Main Proof

Our main goal is to control the size of the set actk defined by Algorithm 1. In turn, these bounds depend on the size of the confidence bands for f*(x) (denoted by δk). Suppose LN is the number of iterations performed by the algorithm before termination.

Let Nactk ≔ ⌊Nk · Π(actk)⌋ be the number of labels requested on the kth iteration of the algorithm. We first claim that the following bounds hold uniformly for all 1 ⩽ kL with probability at least 1 − α: (A.6) f*-f^k, act kC1log(N/α)+dlog(Nk)Nk1/(d+2),Π( act k)C2log(N/α)+dlog(Nk-1)Nk-1γ/(d+2),(A.6) where Cj = Cj(M, c1, c2, L, LK, ‖K, ℓK, RK, γ), j = 1, 2. This claim will be proved later.

Let E be the event of probability ⩾ 1 − α on which both inequalities of (EquationA.6) hold, and assume that it occurs. Second inequality of (EquationA.6) implies, together with the fact that Nk = 2Nk − 1 by definition, that the number of randomized subjects on each step 1 ⩽ kL satisfies Nk act =NkΠ( act k)2Nk-12+d-γ2+dlog(N/α)+dlog(Nk-1)γ/(d+2)with probability ⩾ 1 − α. If N is the maximum number of randomized subjects the algorithm is allowed to request, then Nk=0LNk act 2log(N/α)+dlog(NL)γ/(d+2)k=0LNk2+d-γ2+dC3(γ,d)log(N/α)+dlog(NL)γ/(d+2)NL2+d-γ2+d,and one easily deduces that on the last iteration L we have (A.7) NLc(γ,Π,d)Nlog(N/α)2+d2+d-γ.(A.7) Recall that NL is defined in Algorithm 1.

To obtain the risk bound of the theorem from (EquationA.7), we apply Lemma A.2: (A.8) V(D^)-V(D*)C(γ)(f^L-f*)·Isign(f^L)D*, supp (Π)1+γ.(A.8) Since sign(f^L)D* act L whenever bounds (EquationA.6) hold, it remains to estimate f^L-f*, act L. Recalling the first inequality of (EquationA.6) once again (for k = L), we get (f^L-f*), act LC1log(N/α)+dlog(NL)NL1/(d+2)C˜N-12+d-γlog(N/α)q,where q=4+2d-γ(2+d)(2+d-γ), which together with (EquationA.8) yields the final result.

It remains to show both inequalities of (EquationA.6). We start with the bound on f^k-f*, act k. First, note that by construction, for every k ⩾ 1 the samples (X(i, k), A(i, k), R(i, k)), i = 1…⌊NkΠ(actk)⌋ are conditionally independent given the data i=1k-1Si collected on steps 1, …, k − 1, with conditional distribution of X(i, k) being Π act k. Thus, we can apply Corollary A.1 conditionally on i=1k-1Si with t=log4Nα to get that with probability ⩾ 1 − α/N, f^k-f*, act k4Clogα4N+dlog(NkΠ( act k)/Π( act k))NkΠ( act k)/Π( act k)1/(d+2)8Chk.It remains to integrate the bound with respect to the distribution of i=1k-1Si: Pf^k-f*, act k8Chk=EPf^k-f*, act k8Chk|i=1k-1SiαN.The union bound over all 1 ⩽ kLN gives the result.

Finally, we will prove the second inequality of (EquationA.6), the bound for the size of the active sets actk. This is the place where assumption (A3) on the tree decomposition and margin assumption (A4) play the key role. To obtain the bound, we will compare two estimators of f*: the first is the kernel estimator f^k constructed by the Algorithm 1 on step k, and the second is the piecewise-constant estimator fk with similar approximation properties to f^k. Namely, fk is the L2(Π)-projection of f* on the linear space of piecewise-constant functions of the form g(x)=j=1J(mk)αjI{Tmk,j}(x),αjR. Recall that Ti, j is defined in the tree decomposition of Section 7. As a result, we will be able to relate the “active sets” associated to these estimators, taking advantage of the fact that the active set associated to fk is always a union of the sets from a collection {Tmk,j,j=1...J(mk)}.

Let E1 be the event of probability ⩾ 1 − α on which f^k-f*, act kδk for any k ⩾ 0, where δk = 4Chk. Assume that E1 occurs.

The following inclusions hold (for the definition of ASk+1, see Algorithm 1): (A.9) x:|f*(x)|<δk/2ASk+1x:|f*(x)|<5δk/2.(A.9) Indeed, |f*(x)|<δk/2|f^k(x)|<δk/2+|f*(x)-f^k(x)|<32δkxASk+1and xASk+1|f^k(x)|<32δk|f*(x)|<52δk.For all xTmk,j, set fk(x):=1Π(Tmk,j)Tmk,jf*(y)dΠ(y), and note that |f*(x)-fk(x)|1Π(Tmk,j)Tmk,j|f*(y)-f*(x)|dΠ(y)2LΠ(Tmk,j)Tmk,j|x-y|dΠ(y)2L diam (Tmk,j)2LK12-mk4LK1hk,where the last two inequalities follow from part 3 of assumption (A3) given in Appendix 7, and from the definition of mk. Define τk ≔ max (5δk, 4LK1hk) ⩽ C5δk, Fk+1:={f:|f(x)-fk(x)|(3/2)τk,x act k} to be the band of size (3/2)τk around fk, and Ak+1:=x act k:f1,f2Fk+1,sign(f1(x))sign(f2(x)).

By a reasoning similar to above, we have the inclusions (A.10) x:|f*(x)|<τk/2Ak+1x:|f*(x)|<5τk/2.(A.10) Moreover, by the definition of τk, we have the inequality 5δk/2 ⩽ τk/2. Hence (EquationA.9) and (EquationA.10) imply that ASk+1Ak+1. It remains to note that

1.

Ak+1 is the union of the sets from a collection {Tmk,j,j=1...J(mk)}, hence Ak+1 act k+1;

2.

By (EquationA.10) and assumption (A4), Π( act k+1)Π(Ak+1)Π(x:|f*(x)|<5τk/2)K2(5τk/2)γC6δkγ,hence proving the claim.

Supplementary Materials

Supplementary materials available online include: additional simulation results with normal biomarkers; sensitivity analysis regarding different prespecified parameters; performance of a doubly robust augmented inverse probability weighted estimator; and proof of Lemma A.1.

Acknowledgment

Guang Cheng was visiting SAMSI and on sabbatical at Princeton while this work was carried out and revised; he would like to thank SAMSI and Princeton ORFE department for their hospitality and support. We thank Dr. A. John Rush and the investigators for use of their data from the Nefazodone CBASP trial. The stimulant data used in this article were obtained from the datasets distributed by the NIDA.

Funding

Research Sponsored by NSF (DMS-0906497, CAREER Award DMS-1151692, DMS-1418042), Simons Fellowship in Mathematics, Office of Naval Research (ONR 11845813), and a grant from Indiana Clinical and Translational Sciences Institute.

Log in via your institution

Log in to Taylor & Francis Online

PDF download + Online access

  • 48 hours access to article PDF & online version
  • Article PDF can be downloaded
  • Article PDF can be printed
USD 61.00 Add to cart

Issue Purchase

  • 30 days online access to complete issue
  • Article PDFs can be downloaded
  • Article PDFs can be printed
USD 343.00 Add to cart

* Local tax will be added as applicable

Related Research

People also read lists articles that other readers of this article have read.

Recommended articles lists articles that we recommend and is powered by our AI driven recommendation engine.

Cited by lists all citing articles based on Crossref citations.
Articles with the Crossref icon will open in a new tab.