if - Department of Statistics
Transcription
if - Department of Statistics
.. • GROUP SEQUENTIAL METHODS FOR BIVARIATE SURVIVAL DATA IN CLINICAL TRIALS: A PROPOSED ANALYTIC METHOD by Sergio R. Munoz Department of Biostatistics University of North Carolina • Institute of Statistics Mimeo Series No. 2140T December 1994 • GROUP SEQUENTIAL METHODS FOR BIVARIATE SURVIVAL DATA IN CLINICAL TRIALS A PROPOSED ANALYTIC METHOD. Sergio R. Muiloz A dissertation submitted to the faculty of the UniYersity of North Carolina in partial fulfillment of the requirements for the degree of Doctor of Philosophy in the Department of Biostatistics. Chapel Hill 1994 ApproYed by: J4W-~,~ (b2 2:: ~ E.a: /dYiSOr ~ a~ 1\. C~~r- CO Adv's", Reader Reader Reader ABSTRACT Sergio R. Munoz. Group Sequential Methods for Bivariate Survival Data in Clinical Trials: A Proposed Analytic Method. (Under the direction of Shrikant I. Bangdiwala and Pranab K. Sen) Interim analysis on accumulating data in clinical trials is usually done to assess whether there are significant differences between the experimental treatment under study and the control treatment in order to decide whether or not to stop the trial prematurely. Among many reasons for doing interim analysis are the possible early evidence of treatment differences and also the ethical considerations that subjects should not be exposed to an unsafe, inferior or ineffective treatment. Group sequential designs (Fleming & DeMets, 1993) are suited for doing interim analysis as they allow correction of the type I error which is known to increase as a consequence of repeated testing on accumulated data. In many studies, more than one observation is obtained from the same individual, e.g. studies on women where both breasts are the units of analysis, odontological studies where several teeth are considered, ophthalmological studies where both eyes are analyzed. In such situations it is necessary to account for the correlated structure in the data in the context of group sequential analyses. A model reflecting the correlated nature of a bivariate parametric survival distribution is developed for the case of the bivariate exponential distribution of Sarkar (1987). The uncensored and censored type of models are presented. The model incorporates information from each of the individual organs, which is more efficient than using information from subjects based on a single outcome summarizing the failure of the related organs. The developed model is presented and justified on statistical and biological grounds. The methodology of group sequential testing is developed after parameter estimation and construction of suitable test statistics. Numerical simulations are used to illustrate the application of the methodology developed. . ACKNOWLEDGEMENTS I would like to thank my committee members, Drs. Shrikant I. Bangdiwala, Pranab K. Sen, Harry A. Guess, Clarence E. Davis and Jianwen Cai for their increasing support and thoughtful comments and suggestions. In addition, I would like to recognize Dr. Bangdiwala (Kant) for being exponentially biased to friendly and judicial advice. To all of them "Many Thanks". I also gratefully acknowledge the International Clinical Epidemiology Network (INCLEN) and the "Study Grant Program of the W.K. Kellogg Foundation" for allowed me to complete my doctoral studies and research. Last but not least, I want to recognize the enormous amount of love, understanding and sacrifice my family have done in my favor. My degree is a result of their faith in me, and I want to dedicate it to them and to the memory of my brother Fernando. 11 TABLE OF CONTENTS Page v LIST OF TABLES LIST OF FIGURES VI Chapter 1. INTRODUCTION 1 1.1 Motivation 1 1.2 Brief Summary of work done in this research 2 2 LITERATURE REVIEW AND MODEL PRESENTATION 5 2.1 Group Sequential Analysis 5 2.1.1 Introduction 5 2.1.2 Haybittle-Peto Boundaries 9 2.1.3 Pocock Boundaries 11 2.1.4 O'Brien-Fleming Boundaries 13 2.1.5 Lan and DeMets Boundaries 20 2.1.6 Bayesian Boundaries 22 2.2 Group Sequential Analysis for Survival Data 25 2.2.1 Introduction 25 2.2.2 Log-Rank Test and the Efficient Score Test for the Proportional Hazard Model 25 2.2.3 Boundaries for Group Sequential Tests 31 2.3 Analysis of Clustered Data 33 3 MODEL FOR UNCENSORED DISTRIBUTION OF SARKAR BIVARIATE EXPONENTIAL 38 3.1 Introduction 38 3.2 Vector of Bivariate Hazard Rate 38 III . • 3.2.1 Definition of Vector of Multivariate Hazard Rate 38 3.3 Sarkar's exponetial Distribution 41 3.3.1 Introduction 41 3.3.2 Definition and Properties of the Distribution 45 3.3.2.1 Definition 45 3.3.2.2 Properties 46 3.3.3 Marginal MLE' s of A. and y 47 3.3.4 Joint MLE's of A. and y 48 3.3.5 Estimating Equations and Information Matrix 50 3.3.5.1 Estimating Equations 50 3.3.5.2 Observed Information Matrix 51 4 • Sarkar Bivariate Exponential MODEL FOR CENSORED DISTRIBUTION OF SARKAR BIVARIATE EXPONENTIAL 54 4.1 Censoring Schemes 54 4.2 Model Specification and Maximum likelihood Estimation 55 4.3 Estimating Equations and Information Matrix 60 4.3.1 Estimating Equations 60 4.3.2 Observed Information Matrix 62 5 GROUP SEQUENTIAL TEST 66 5.1 Introduction 66 5.2 Bivariate Group Sequential Test 67 5.3 Concluding Remarks 71 6 NUMERICAL RESULTS 72 6.1 Introduction 72 6.2 Single analysis results st the End of the study 76 6.2.1 Ignoring the bivariate nature of the data 76 6.2.2 Incorporating the bivatiate nature of the data 79 6.3 Group Sequential boundaries 81 7 FUTURE RESEARCH 83 References 87 IV LIST OF TABLES Page Table 8 2.1 Decisional Procedure in Group Sequential Test 2.2 Pocok Boundaries for a Two-Sided Group Sequential Test with Type I error a 12 O'Brien-Fleming Boundaries for a Two-Sided Group Sequential Test with Type I error a 14 Haybittle-Peto, Pocock and O'Brien-Fleming Boundaries for a Two-Sided Group Sequential Test with Type I error a=0.05 and number of tests K=4 16 2.3 2.4 2.5 Group Sequential Boundaries for Haybittle-Peto, Pocock, O'Brien-Fleming and Bayesian Approaches for 200 subjects 2.6 Number of Events at Tiem Tk for subjects at risk, by treatment group 6.1 Simulation Scenarios and values for sample of size 500: Bivariate Exponential distribution of Sarkar 23 27 74 6.2 Summary statistics for scenarios A, B, C 78 6.3 Test statistic for the biavariate exponential distribution of Sarkar 80 lr v LIST OF FIGURES Figure , 2.1 2.2 Page Habittle-Peto, Pocock and O'Brien-Fleming Critical Values for a Four Group Sequential Test at 0.05 Significance Level Habittle-Peto, Pocock and O'Brien-Fleming Critical Values for a Nine Group Sequential Test at 0.05 Significance Level 18 19 6.1 Simulated Bivariate Exponential Distribution of Sarkar 75 6.2 Kaplan-Meier survival curves for scenario A 77 VI CHAPTER 1 Introduction ·, 1.1.-Motivation In analyzing clinical trials, there are many statistical aspects to be considered. The type of endpoint to be analyzed, the sampling design and repeated looks at the data for monitoring purposes, are, among others, considerations in doing statistical analysis. Regarding these three considerations, our mam interest repeated looks at correlated bivariate time to-event-data. It IS focused in doing IS well known that observations being taken from the same subject tend to be more alike than measurements being gathered from independent observations. Ignoring the correlation structure and • treating correlated data as if they were independent can result in misleading analyses depending on the magnitude of the intracluster correlation. There are several examples of this type of data, including studies on women where both breasts are the units of analysis, odontological studies where several teeth are considered, ophthalmologic studies where both eyes are analyzed, and other situations that convey to naturally clustered data. Community studies where a sample of schools, villages, or workplaces are gathered are also a very important source of clustered data. In all these situations, for practical or feasibility considerations, the investigator intervenes on the cluster (person, school, community), but would prefer to evaluate the endpoint in each of the individual members of the cluster. Aside from power considerations, this may also be the logical choice for unit of analysis. On the other hand, clinical trials are designed with ethical considerations for the subject and usually thus involve interim analysis of data prior to the scheduled end of the data collection period. Interim analysis on accumulating data in clinical trials is usually done to assess whether there are significant differences between the experimental , treatment under study and the control treatment in order to essentially decide whether or not to stop the trial prematurely. Among the many reasons for doing interim analysis are the possible early evidence of treatment differences and also the ethical considerations that subjects should not be exposed to an unsafe, inferior or ineffective treatment (Jennison and Turnbull, 1991). Group sequential designs (DeMets, 1987) are suited for doing interim analysis as they allow correction of the type I error which is known to increase as a consequence of repeated testing on cumulated data. Several procedures have been proposed in the last fifteen years and some of them are going to be discussed in a later chapter. . Most of the procedures for doing interim analyses are suited to analyze independent observations; therefore, adjustment for clustered data would need to be done. The purpose of this dissertation is to propose a methodology that allows us to analyze clustered bivariate survival data under the group sequential framework. 1.2.- Brief summary of work done in this research. This dissertation intends to provide a solution for solving the problem of doing interim analysis when parametric dependent bivariate survival data is gathered from each subject, taking into account the correlated nature of the data and the needed adjustment for a fixed overall type I error probability. 2 The specific model to be studied deals with subjects contributing two organs to the analysis such as two eyes, two kidneys or two breasts. A real life example of this type of data is taken from the "Sorbinil retinopathy trials research, 1990" . They . evaluated the drug Sorbinil, an aldose reluctance inhibitor, for ocular diabetic retinopathy on a sample of 497 patients with insulin dependent diabetes mellitus. Patients were randomly assigned to take either oral Sorbinil or placebo and the occurrence of severe visual loss was one of the outcomes being analyzed. In this case, patients are contributing two failure time observations to the analysis. These failure times may be correlated and possibly censored. The goal of the proposed research is to obtain adjusted interim analysis boundaries for clusters of size two when K looks at the data are planned. The model considers the joint parametric bivariate exponential distribution of Sarkar (1987). Each component of the vector is based on each of the individual organs of a single patient, which is more efficient than using a hazard based on a single outcome summarizing the failure of the related organs of the patient. This joint distribution is assumed to be symmetric, so that the corresponding marginal hazard function of each organ is the same. Furthermore, the model considers the hazard function to be altered for the remaining organ once the failure of one of them occurs. In other words, it is assumed that patients have the same marginal hazard function for each organ while no failure has occurred, but after one of the two organs fails, the hazard function of the remaining organ is now different from the original marginal hazard function for that particular organ. Thus, the remaining organ's hazard function is a new hazard conditional on several characteristics determined by the failure of the first organ. 3 Chapter 2 contains the literature review of group sequential tests for non censored and censored data; it also contains a review of methodology on analysis of clustered data. In chapter 3, the bivariate exponential distribution of Sarkar is examined under no censoring and maximum likelihood estimators and variances are derived. Censoring is incorporated in chapter 4. Chapter 5 contains the proposed group sequential test statistics. Numerical results are presented in chapter 6 to illustrate the methodology. Concluding remarks and future research are presented in chapter 7. 4 . CHAPTER 2 Literature review and Model Presentation 2.1.- Group sequential analysis 2.1.1.- Introduction Interim analysis is defined as any assessment of data done during either patient enrollment or follow-up stages of the trial for the main purpose (among others) of assessing treatment effects. Sometimes interim analysis conveys the decision of stopping the trial. If the trial is stopped early for concern that the .. experimental treatment may increase the incidence of death, then there is no more data considered, and there are no subsequent statistical inferences considered beyond estimation of the incidence rate. The complications caused by interim analyses upon the estimation problem are not being considered in this dissertation. However, when the trial is not stopped early, standard hypothesis testing and confidence intervals of the treatment effects do require adjustment for the previous analyses and are the topic of this dissertation. Table 1 shows schematically what the decision procedure is in group sequential trials. Traditional methods (fixed sample size) for the calculation of significance levels and confidence intervals are not valid after repeated interim analyses, because they do not achieve the error rates required. Multiple significance tests can greatly increase the chance of false positive findings. The best known statistical instrument for interim analyses in controlled clinical trials is group sequential methodology. These methods depend crucially on the knowledge of the joint distribution of the test statistic at the monitoring time. To illustrate the general procedure of group sequential analysis, let us assume that an experimental treatment is being compared against a control treatment and a planned total of N patients is divided into K groups of 2n patients each (2nK=N). Assume now that a group of 2n patients are randomly allocated to each treatment so that n patients are assigned to the experimental treatment and n are assigned to the control treatment. The decision to stop the trial or continue is based on repeated significance tests of the accumulated data after each group is evaluated. In real life problems, this fact is very restrictive since data is continuously gathered. Another restriction is that it is also assumed that response to treatments being compared is immediate. Let Zi represent the test statistic for the i-th group and ci the critical value associated with Zi ' i ::; K. Starting at the first interim analysis, in testing the hypothesis of no treatment difference, the decision is stop the trial if IZ 11 > c1 or take another group of 2n patients and randomly allocate them into the treatments in the same way described above. If the trial was not stopped at the previous stage, compute Z2 based on the 4n patients and stop the trial if IZ21 > c2 or take a new group of 2n patients to be allocated into the treatment being tested. Keep doing the same procedure until either you stop the trial at any of the stages or take a new group of patients up to the last group. The values of ci are chosen such that Pr{ZI>CJ or Z2 >c2 or ....or ZK>cK I HO}=a., 6 where a is the pre-specified significance level to test the null hypothesis of no treatment effect difference. One point to notice so far is that the number of groups has to be specified in advance. Table 1 illustrates the procedure. . 7 Table 2.1: Decisional Procedure in Group Sequential Trials. Condition Gro up Sample size Decision 1 2n ar Stop the trial IZll > cl br Take a 2nd group IZII:S; cl a2: Stop the trial IZ21 > c2 b2: Take a 3-th group IZll:S; cl; IZ21:S; c2 ar Stop the trial IZil > ci bi: Take a i-th group IZll:S; cl ·.. ·... IZi_ll:S; ci-l" IZil:S; ci aK RejeetHo IZKI>CK hi( Do not reject Ho IZll:S; cl ····.·.IZK-ll:S; CK-l··IZKI:s; CK " 2 i K 4n 2in 2nK=N " • 8 Several authors have proposed a variety of criteria and of methods for • choosing the critical values ci, to address the issue of requiring a total probability of type I error of a pre-specified. Some of the most common strategies are presented below. 2.1.2.- Haybittle-Peto Boundaries. Haybittle (1971) and later Peto (1976) defined ad-hoc boundaries for group sequential tests so that CJ = c2 = ....= cK-1 were chosen in such a way that the total probability of crossing these boundaries were almost negligible and finally adjusting cK, so that the overall type I error probability was a. Although .. it may appear pointless to set up a sequential test in which the probability of early stopping is almost negligible under the null hypothesis, the approach does provide reasonable power to detect the kind of major differences in efficacy which would make early stopping an ethical necessity. Consider the particular case of comparing the mean of an experimental treatment against a control treatment and assume that XE and Xc represent the experimental treatment and the control treatment responses, respectively. Further • assume that both XE and Xc are normally distributed with known common variance 0 2 and mean J..lE and J..lC, respectively. Assume that the null hypothesis is that of no treatment difference against the alternative hypothesis of treatment difference and that a total probability of type I error equal to a is desired. 9 The goal is to compare K consecutive groups of 2n individuals where n of them are randomly assigned to each treatment. Denote by X Ek and XCk the observed mean responses of treatment and control group at the k-th interim If analysis (k=1,2, .... ,K). l t(X xc) Define the statistic d k = k j=1 E J for k=1,2, ... K. Then dk is J normally distributed with mean equal to IlE - IlC and variance equal to 2cr2/nk. Notice that the increment dk-dk-l is statistically independent of dk-l and it is normally distributed with mean IlE - IlC and variance equal to 2cr2/n. To test the null hypothesis of no treatment effect difference versus the two-sided alternative that experimental treatment effect is different from the • control one at a significance level of ct, the statistic to be used is dk.Jnk Zk = ~ N(O,l). . .J2cr The Haybittle -Peto procedure is to reject the hypothesis of no treatment difference if Idkl ~c H_ P k .J~ , vnk for k=l, 2, ...., (K-l) and where cH_P k is a fixed ' constant for every k less or equal to K-l, and finally for k=K, cH-PK , is chosen such that the total probability of Type I error is equal to ct. Equivalently, the rejection region can be written as IZkl ~ cH_P,k In practice, taking for example K=4, and assuming that the statistic to be used to test the null hypothesis of no treatment difference corresponds to a test 10 derived from a standard normal distribution, then c1 = c2 = c3=3.0 and C4=1.98 for a desired overall level a=O.05 . • 2.1.3.- Pocock Boundaries. The name of group sequential design was formally introduced by Pocock (1977). He suggested a constant adjustment to the K fixed critical values associated with the K repeated tests coming from those K planned interim analyses. The decision of rejecting the null hypothesis of no treatment difference is • based on the same statistic Zk as defined above. The critical values ck are chosen such that the null hypothesis is rejected after the k-th group if IZkl > cp(k)=cp, .. independent of k, for k= 1, 2, ..... , K, where cp is a constant value chosen such that the overall type I error probability is a. Notice that cp values are constant for every k=1, 2, ...., K; they are different from the constant fixed values proposed by Haybittle and Peto for the K-l first groups. As an illustration of the Pocock procedure, consider the .two mean comparison introduced in the previous section. The procedure is to reject the hypothesis of no treatment difference if Idkl • ~ c .J~ vnk p for k=1, 2, ....., K, and where cp is a constant value. Pocock provided cp values for different values of K and levels of significance. Table 2 reproduces some of those values. 11 Table 2.2 * Pocock cp Boundaries for a two sided Group Sequential Tests with type I error u. K u=0.05 u=0.10 2 2.178 1.875 3 2.289 1.992 4 2.361 2.067 5 2.413 2.122 2.555 • Extracted from Pocock ,1971. • 2.270 10 .. 12 In the case of K=4 and cx=0.05, the decision of rejecting the null hypothesis is based on the fixed critical value 2.361 for every interim analysis. This compares with the Haybittle-Peto critical value of 3.0 for k=1,2, and 3 and 1.98 for the final analysis (K=4). 2.1.4.- O'Brien - Fleming Boundaries. The O'Brien - Fleming procedure is also based on the assumption that K, the number of interim analyses, is predetermined and that groups of 2n individuals are randomly assigned to the experimental and control groups respectively. The critical values proposed by O'Brien - Fleming (1979) .. monotonically decrease with k. This is intuitively accepted since one does not wish to stop in the early stages of the trial unless differences are substantial between the groups being compared. For the comparison of two means presented earlier, the rejection region of the no treatment difference null hypothesis is given by values of dk such that I Co-F Id k ~.Jk * ~2d' .. k .Jnk' whic. h'IS decreasmg m . O'Brien - Fleming provided simulated cO-F values for different values of • K, the number of interim analyses, and significance level cx. Table 3 reproduces some of those values. 13 Table 2.3: * O'Brien - Fleming cO-F Boundaries for a two sided Group Sequential Tests with type I error u. K u=0.05 u=0.10 2 2.797 2.373 3 3.471 2.962 4 4.048 3.466 5 4.562 3.915 6.597 10 5.595 , Extracted from O'Bnen - FIeIDlng, 1979 14 Table 4 and Figure 1 compare the Haybittle - Peto, Pocock and O'Brien Fleming procedures of group sequential test critical values with K=4 pre-planned interim analyses and a=O.05. Figure 2 shows a similar comparison for K=9. Figures present only the upper region since the lower one is a symmetric copy of the upper one. . 15 Table 2.4: Haybittle-Peto, Pocock and O'Brien - Fleming Boundaries for a two sided Group Sequential Tests with type I error a=0.05, number of tests K=4. Critical Value Haybittle-Peto Pocock O'Brien-Fleming cI 3.0 2.36 4.05 c2 3.0 2.36 2.86 c3 3.0 2.36 2.34 c4 1.98 2.36 2.02 16 From figure 2 we can see that the O'Brien-Fleming's method is more conservative at the beginning of the trial when it is compared with both Haybittle-Peto and Pocock boundaries. In doing either four or nine interim analyses, the probability of rejecting the null hypothesis at earlier stages of the trial is smaller for O'Brien-Fleming method, but the situation is reversed at latest stages of the trial. ... 17 Figure 1. Haybittle-Peto, Pocock and O'Brien-Fleming absolute critical values for a four group Sequential tets at 0.05 significance level. 4.5 . , . . - - - - - - - - - - - - - - - - - - - - - - - , 4 A~ •• 3.5 3 ~ ~ •• - •• - : .. ~.t 2.5 •• 2 -. .----- .-.... ...... ..... . .. .. .. - . . . _ _ -·.::..:..~. . :-;:-~ ........- :-II •••• - """0 - • - Haybittle-Peto - . -Pocock 1.5 • •• • O'Brien-fleming 1 0.5 o+--------+--..;..-----+-------~ 1 3 2 number of interim analyses (K). . 18 Figure 2. Haybittle-Peto, Pocock and O'Brien-Fleming critical values for a nine group Sequential tests ata 0.05 significance level. .. 7-r-----------------------, 5 4 - . -Pocock - • • • O'Brien Fleming - . . . . Hayblttle-Peto 1 O+--~I----+--_+_--+_-_t--_+--+_---f 1 2 3 4 5 6 7 number of interim analyses (K)• .. 19 8 9 2.1.5.- Lan and DeMets Boundaries. Two main requirements of the group sequential procedures proposed by Haybittle-Peto, Pocock and O'Brien-Fleming are (a) previous specification of the number of interim analyses K to be performed and (b) equal increments of information, say 2n observations in each group. This requirement does not conform with the actual practical conduct of clinical trials. Interim analyses are performed at predetermined time periods, and the number of observations in each group at the k-th look is thus a random variable. Lan and DeMets (1983) proposed a procedure to address the two constraints above specified allowing for a total Type I error probability equal to a. They proposed a flexible discrete boundary, called cO_F(k), for the sequence of statistics Zk-, k=1, 2, ..., K. The procedure is based on the choice of a function aCt), called the "spending function", which characterizes the rate at which the Type I error level a is spent. Assuming that completion of the trial by time T is scaled arbitrarily such that T=1, then the function a(t) is built so that a (0)=0 and a( 1)=a. This function gives cumulative probabilities of Type I error and it allocates the amount of Type I error one can "spend" at each analysis. If t represents the first boundary crossing of a standardized test statistics Zk, then aCt) can be written as: aCt) = P{ t ~ t} , o~ t ~ 1. Both Pocock and O'Brien-Fleming boundaries can be approximated (DeMets, 1987) by functions as defined above: (a) al (t)=alog{1+(e-l )t}approximates the Pocock boundaries, and 20 .. (b)a2(t)=2-2<I>(1.96/"t) approximates the O'Brien-fleming boundaries for a =0.05, where <I> denotes the standard normal cumulative distribution function. For K groups, set arbitrary points, 0 < t1 < t2 <.......<tK:::; 1, such that tk is the scaled time for the k-th analysis (k=l, 2, ... K). The Lan and DeMets boundary point cL_o(k) is chosen such that under the null hypothesis of no treatment difference and given cL-O(l), cL_O(2), ........, cL_o(K-1) the probability P{I Z 11::;cL-O(l), IZ 21::;cL_O(2), ..... IZ k-11::;cL_O(k-1) , IZkl>cL-O(k)}= a(tk)-a(tk_ 1), where Zk represents the standardized test statistic at t=tk, k=l, 2, .... ,.K. The increment a(tk)-a(tk-1) represents the additional amount of significance level that one uses at time tk- This procedure implicitly incorporates the concept of "information time". On the scaled [0-1] interval, the information time t represents the fraction of patients randomized up to that time. Having the knowledge of the target sample size, calendar time can be transformed into information time having the knowledge of the target sample size. This issue allows for analysis not having equal sized groups constituting an advantage over the previous methods before described. 21 2.1.6.- Bayesian Boundaries. Bayesian monitoring of clinical trials has been addressed by several authors. (Berry, 1985, 1989; Freedman and Spiegelhalter ,1989; Jennison and Turnbull, 1990) among others have presented the Bayesian approach to group sequential tests as an alternative to the previous methodologies presented above. Consider a randomized clinical trial comparing two treatments. Let 8 denote the true treatment difference, with a pre-trial prior distribution denoted by p(8). Suppose further that n pairs of subjects have been entered into the study and one subject of the pair has been assigned to the experimental treatment and A the other one to the control treatment. Let 5 denotes the estimated treatment difference. Assume that (i) the prior distribution of 8 is given by p(8) 2/no) and (ii) the conditional distribution of the estimate A 8" ~ N(80 ' (j A given 8 is p( 8" 15) ~ N(8, (j2/n ), where (j2 is the variance of an individual pairwise difference, 80 .is the pre-trial expectation of the treatment difference, and no reflects the precision of the prior information about the treatment difference. A ~ The posterior distribution p(8 18 ) is given by: p(8 18 ) i ~ n8+no8o c? ] ----=-.::....; . For a two sided test, assume that 8 > L\T and 8 < L\c n+n o n+ no reflect values where the experimental treatment is considered superior to the control and values where the control treatment is considered to be superior to the experimental one, respectively. The decision whether or not to stop the trial is based on numbers Pc and PE such that one should stop the trial if either: 22 r 8 Pc = p(018)do < E or PE -~ = r p(018)do < E . 8£ In words, the trial is stopped if either there is a very little chance that the control treatment is superior (experimental treatment is preferred) or the same is true for . the experimental treatment (control treatment is better). Since the posterior distribution of 0 is normal, rejection conditions can be rewritten as follows: (i) "8 > c? n i< c?[~E _oono _<1>-I(E)Jl lJ n ~ c? , l' ~c~ - oono - <1>-1 (E) lJ c? and (ii) op where <1>- 1 (8) is the inverse of the standard op normal cumulative distribution at point E, and o~ 2 = 0 n+no is the variance of the posterior distribution of o. As one can notice, Bayesian boundaries are dependent on the choice of the prior and mainly on the choice of the prior variance. Table 5 shows three Bayesian boundaries, assuming 0 2 =0.05, 00=0 and E =0.025, and their comparison with Haybittle-Peto, Pocock and O'Brien-Fleming boundaries on the basis of 200 subjects for testing the null hypothesis of 0=0. The three Bayesian rules correspond to three different standard deviations (o/-J no) of the prior distribution, namely 0.025, 0.075, and 0.15. . 23 Table 2.5. Group sequential boundaries· for Haybittle-Peto, Pocock, O'BrienFleming and Bayesian approaches for 200 subjects. .. !:i. Non - Bayesian Bayesian cr/..}no Interim Analysis # of pairs of patients H-P Pocock O-F 1 20 0.38 0.72 2 40 0.27 3 60 4 5 0.025 0.075 0.15 0.52 0.72 0.45 0.36 0.37 0.39 0.27 0.22 0.24 0.30 0.68 0.28 0.21 80 0.19 0.18 0.26 0.51 0.23 0.18 100 0.17 0.14 0.14 0.42 0.19 0.15 * Extracted from Friedman and Spiegelhalter (1989) 24 2.2.- Group sequential analysis for survival data. 2.2.1.- Introduction One of the most frequently used test statistics in the analysis of survival data in clinical trials is the log-rank test, where patients entering the study are assigned randomly to the treatments. Most of these trials are designed so that the log-rank test is evaluated at the end of the study. The log-rank test for group sequential analysis was proposed by Tsiatis (1981). His derivation is based on the efficient score test for the proportional hazard model, and shows that the logrank test is the special case when treatment assignment is the only covariate in the model. The asymptotic joint distribution of the efficient score test for the proportional hazard model at different points in time was established to be used for repeated significance tests. 2.2.• 2.- Log-rank test and the efficient score test for the proportional hazard model. The development of the log rank test, Mantel, (1966) and Peto (1972); to compare two treatments uses a conditioning argument based on the number of subjects at risk of having the event just prior to each observed event time. Let T1 < T2 < < TL denote the ordered observed distinct event time points in the sample formed by combining the two treatment groups to be compared, and let Eik and Aik , k=1, 2, .....L, denote the number of observed events and the number of subjects at risk, respectively, in group i at time point 25 Tk Let Ek and Ak denote the corresponding values in the combined sample. The data at time Tk can be represented as in table 6. 26 Table 2.6. Number of events and no events at time Tk from those subjects at risk, by treatment group. Treatment Group Event Experimental Control Total Yes Elk E2k Ek No Alk - Elk A2k - E2k Ak-Ek Total Alk A2k Ak 27 Given Aik' the Eik have a binomial distribution with number of trials Aik and, under the null hypothesis of a common event rate 'A in the two treatment groups, approximate event probability 'A(Tk)~t. Fisher's exact test for equal binomial parameters (experimental and control treatment groups) is based on conditioning further on Ek. Given the above stated conditions, Elk has a hypergeometric distribution with mean and variance given by: Alk [] A lkA 2k Ak- E k -2 Elk = E [Elk ] = Ek --- ; ulk = V Elk = E k Ak A Ak- 1 k Given the margins in each of the L tables at the observed event times, the vector {Ell - Ell> E 12 - E12' ....... , ElL - Eld is a vector of observed minus conditionally expected number of event across observed event times. Further, assuming independence among the elements of this vector, then the statistic defined by: known as the standardized two sample log rank test, has approximately standard normal distribution. Let us now introduce the Cox proportional hazard model (Cox, 1972). Let 'A(t;x) represents the hazard function at time t for a subject with the set of covariates x=(xl> x2, ...., xp)'. The Cox proportional hazard model specifies that: 'A(t;x) = 'Ao(t)ex~, where "-o(t) is an arbitrary unspecified base-line hazard function (all covariates are set equal to zero) and J3'=(~l> ~2, parameters. 28 ....., ~p) is a vector of regression .. The partial likelihood for P is formed taking the product over all points ti, i=l, 2, ...., k, to produce: L L(~) = nJ exp(xi~) • i=l l LexP(x1P)J Dei, i=l leA(t i ) where 8 i represents the conditional probability that subject i has the event at time ti given that the total number of subjects at risk at that time is A(ti) and that exactly one event occurs at ti. When comparing the survival time subject to censoring among two groups without any other covariate, and x denoting an indicator function with value one if experimental treatment and zero if control, the corresponding hazard function and partial likelihood related to the Cox proportional hazard model are expressed as follows: A(t;X)=AO(t)e x ~ and L(P)=D J Lexp(xtP) exp(x.~) } =D8 k k l j o leA(tj) The log likelihood is: l = Ln(L(~» ~ Xi~ - {~~P(XI~»)} ~Lnei . = { = Since Xi is either zero or one, the log-likelihood can be rewritten as: k f oc P~dli - k ( '1 ~1'te~~xP(Pdlil)) where d Ii is the number of events at time ti in the experimental group and d Iii is the number of subjects at risk of experimental group at time ti. 29 The efficient score statistic and the information for f3 are gIven respectively by: where Aj(f3) and Vj(f3) denote respectively the mean and the variance of d iiI under weighted sampling without replacement from the risk set at time tj. To test the hypothesis of no treatment difference, f3=0, the score test can be applied. The score test requires the computation of S(O) and 1(0). Under the null hypothesis, the sampling scheme defining Alf3) and Vj(f3) reduces to simple random sample without replacement, so that Aj(O) and Vj(O) are just the mean and variance of an hypergeometric random variable and their expressions are: Aj (0) = di rIj rj Vj(O) = dj rOjr~i(ri -d i ) ri (rj -1) where rOj and rIi are the risk set sizes in the control and experimental group respectively. The score test, which compares U(O) to its estimated vanance 1(0) corresponds to the log rank test previously described here. The name comes from the relation to the exponential ordered scores of the test (Cox and Oakes, 1984). This test can also be obtained by setting up a separate 2x2 contingency table at each event time and carrying out the combined Mantel-Haenszel test (piece-wise exponential fitting). 30 • 2.2.3.- Boundaries for Group Sequential Tests. Group sequential boundaries with time to event data are based on the fact that the log-rank test behaves like a partial sum of independent normal random variables with variance proportional to the number of observed events. Therefore, all the group sequential methods described before are applicable, using the standardized two sample log-rank test Q defined previously. Suppose that at calendar time Tc' there are e distinct event times t 1> t2' ....., te, where tj represents the time from entering to the study to the occurrence of the event. Then the Log-rank tests is to be computed a scheduled maximum of N calendar times and at each time Tn, the log-rank test is computed and compared to a boundary cn' n=l, 2, ..., N. The Haybittle-Peto, Pocock and O'Brien-Fleming boundaries can be used and they required the number of looks to be predetermined and to have equal number of events between consecutive looks. This is an equivalent requi~ement for normal and binomial data of equally spaced analyses (equal sized groups). The Lan and DeMets boundaries can be computed using the "estimated information fraction" (Lan, Lachin; 1990). That is estimating the fraction of patients having the event at the time of the analysis. The actual value of this fraction is generally unknown since the total number of events at the end of the trial is unknown. Lan, Reboussin and DeMets (1993) defined what they called 31 the "natural estimate of information" as a function of the cumulative distribution function of the time to event random variable. • . 32 2.3. Analysis of Clustered Data. . Some studies involve randomization of clusters of units rather than units themselves in the allocation of treatment groups. Examples of cluster allocation ranges from studies where both eyes of an individual are allocated to a systemic treatment regimen in ophthalmologic studies; to the evaluation of educational strategies, where either schools or classrooms are randomly allocated to different teaching methods; and to evaluation of health care procedures where medical practices or even entire communities are randomly assigned to either intervention or control group. Application of methods for this kind of data can be found widely in the literature. Donner (1989) and Rosner (1984) proposed statistical methods for analysis of clustered data in ophtalmological studies; DeRouen et al (1991) and Hujoel et al (1991) presented methods applied to dental studies. In both ophtalmological and dental studies, the patient represents the cluster and the eyes or teeth, respectively, represent the units of analysis within clusters. Since units within clusters are generally not statistically independent, randomization of clusters are in most cases less efficient than studies in which the randomization has been carried out on units themselves. Despite this disadvantage, cluster randomization often conveys economical, ethical or administrative practical advantages. In some cases, it is the only feasible allocation procedure available to the investigator. A key issue in cluster randomization is the fact that the variability among clusters tends to be higher than the variability within clusters and the larger the 33 ratio between these two sources of variation, the larger the degree of the within cluster dependency (Donner at aI, 1990). As an example, consider the simple case of sampling one cluster of size two from an infinite population of such clusters. Suppose additionally that a variable y is being measured with expected value Il, variance 0 2 and correlation p between units within clusters. Let Yl and Y2 be the two observations within the cluster. Then, an unbiased estimator of Il is ~ = Y1 + Y2 . 2 If we ignore the fact that Yl and Y2 are correlated, aSSUmIng independence among the observations, the variance of the above proposed unbiased estimator is Vi [~] the dependence 2 = ~. 2 between the On the other hand, the actual variance when observations is taken into account is 2 Vc[~] = ~(1 + p). Therefore, the larger the dependence is among units, the less 2 . efficient is the estimator. The ratio of Vc[~] to Vi[~]is the so called "Kish design effect", denoted by deff(~) = Vc[[~]]. In this example, if p=0.6, then deff(~)=1.6, implying that Vi Il the actual variance of the estimator is 60% larger that what would have being obtained from a sample of the same number of independent observations. Adjustments for dependency have been proposed by several authors and the type of adjustment varies according to the type of data being analyzed. For univariate analysis, Kupper and Haseman (1978) proposed the "Correlated Binomial Model" when the dependency is on dichotomous variables. The model 34 is based on a "correction factor" proposed initially by Bahadur (1961) based on the correction factor applied to the Binomial distribution. They take the second order approximation as the correction factor to the Binomial probability density function ending with the following pdf: P(X ~ x;) =(::";(1-8)n,-,,{ 1+Z8(;-8) [(x; - no)' +X;(Z8-1)-n;8'j} An application of this method to the evaluation of a diagnostic test i n the presence of dependent data can be found in DeRouen et al (1991). Donner (1989) proposed a method he called the "adjusted chi-square test" for comparing proportions coming from sampling clusters of size two. The method is based on an adjustment made to the standard chi-square test for the • equality of two proportions. He considers the fact that the variance of a Bernoulli random variable, with mean 8, is not larger than 8(1-8); therefore, the variance can be approximated, when p:t:O, by the quantity p8(1-8), where p is the intraclass correlation which it is assumed ranges from zero to one. If p=O, then the standard chi-square test is applicable (no dependence among units within the cluster). When p>O, the standard chi-square test statistic can no longer be approximated by a chi-square distribution, but it can be adjusted so that a modified version of the test statistic follows a chi-square distribution. The procedure implies the estimation of p, which can be estimated by the standard analysis of variance estimator of an intraclass correlation for a stratified cluster design. Donald and Donner (1987) proposed adjustment to the Mantel-Haenszel chi-square statistic and to the variance of the Mantel-Haenszel estimate of the common odds ratio in sets of 2x2 contingency tables. They used the same 35 technique described above of approximating the vanance of the binomial distribution by the quantity p8( 1-8). In the case of multivariate analysis, the vanance component model . strategy can be used to account for the dependency of the data incorporating cluster specific covariates as random effects into the model. Sterne et al (1988) applied a variance component model to the study of periodontal disease in which the response is measured at five different sites (teeth) in the same subject. The model allow for components of variation both between and within subjects and therefore avoids the assumption of independence between different sites in the same subject. The only one assumption about the structure of correlations is that the within subject error variance is the same for all sites and subjects. A more general class of models, allowing for a more general correlation structure, is the mixed model, and a recent discussion can be found in Helms (1992). In the previous example, one might allow for a different site-level variance for each type of site; or more generally, a coefficient at either the subject or site-level may be assumed to be random. The generalized estimating equation (GEE) approach (Liang and Zeger, 1986; Zeger and Liang, 1986) is a method which offers the opportunity to apply - regression models for continuous and categorical responses to situations in which observations have dependencies of varying forms. DeRouen et al (1991) applied GEE models to periodontal diseases, where the patient is the cluster and teeth site are the units of analysis within clusters. 36 • For censored survival data, Lee et al (1992) showed that for a large number of small groups of correlated failure time observations, the standard . maximum likelihood estimate of the regression coefficients are still consistent and asymptotically normally distributed; but their variance covariance matrix may no longer be valid due to the dependency among groups. They finally proposed a "correct" variance-covariance estimate which takes account of the intra group correlation. All these methods take into account the dependency among observations in terms of p. What we intend to do for solving our problem is similar, using the vector of bivariate hazard rate, which is described in the next section. • 37 CHAPTER 3 Model for Uncensored Bivariate Exponential Distribution of Sarkar ... 3.1.- Introduction In this dissertation, the problem of accounting for clustered correlated time to failure data is tackled in a parametric setup for the bivariate situation. In this chapter, we initially introduce the concept of bivariate hazard rate and then we present specifically the bivariate exponential distribution of Sarkar. After this, we develop the simple case when the data is complete, that is when there is no censoring. Maximum likelihood estimators and estimated varainces of the estimators are presented under this framework. 3.2.- Vector of Bivariate Hazard Rate. 3.2.1.- Definition of Vector of Multivariate Hazard Rate. Johnson, N.L. and Kotz, S. (1975) defined what they called the joint multivariate hazard rate (JMHR) of m absolutely continuous random variables X J, X2, ....., Xm as the vector: • where Gx(x)=P{Xi>xi, i=l, 2, .... , m}. Thej-th element of the JMHR will be denoted as hx(x)j ( a 'I . =iaxJLnGx(X), J = 1,2, .... ,m. From now on, we will focus on the case of m=2, the joint bivariate hazard rate (JBHR). Some of the properties of the JBHR are: (i).- IfXJ, X2 are mutually independent then hx(x)j = h Xj (Xj) where the left hand side is the j-th component (j=1,2) of a bivariate hazard rate and the right hand side is a univariate hazard rate. (ii). - If the bivariate hazard rate is constant, so that h X (x) = c, which implies that a 'I j ax LnGx(x) = -c j 0=1,2) , then j (2 G x (x) 'I = exp(-cjx)gj(X") (j:;tj'; j=l, 2), therefore G x (x) ex: exol- LCjXj)' • \. j=l J Then the X's are mutually independent exponential random variables if and only if the bivariate hazard rate is constant. (iii).- Noting that Gx(x) = p{X i > Xi; i = 1,2} = p{X 2 > X 2 } *p{X I > XII X 2 > X 2 } =GX2,(X2)*GXIIX2(Xllx2) we can see that: 39 A similar expression can be derived for the 2nd component of the vector. Thus, the components of the vector of bivariate hazard rate are in fact univariate hazard rates of conditional distributions of each variable, given certain inequalities on the remainder. Note that, the first component of the JBHR is: h x X (xJ, x 2h = -~LnGx X (xJ, x 2), which can be rewritten as a function of the I' Oxl 2 I' 2 joint cumulative distribution function and the joint survival function as follows: h x102 x (X1,X2 )1 =- G * :... 1 UA: Xl.X2 (xl.x2) =- G 1 X!.X2 (Xl ,x2) 1 {l-Fx1 (x1)-Fx2 (x 2 )+Fx 102 x (X 1,X2 )} {-fx,(XI )+: !Fx,,x,(xl ,x2 )1} 1 The second component of the vector can be easily obtained by interchanging the indexes in the above expression. Now, given these relationships, we will focus our methodological development on the joint bivariate density function rather than in the joint bivariate hazard rate. 40 .. 3.3.- Sarkar's Bivariate Exponential Distribution (1987) 3.3.1.- Introduction Several bivariate exponential distributions were previously examined in order to detennine the "best" according to both their underlying assumptions and their applicability to real life data. The bivariate exponential distributions we considered initially were the Morgenstern-Gumbel-Farlie Distribution, Gumbel's Bivariate Exponential Distribution and Bivariate Exponential Distributions of Marshall and Olkin. A brief representation of these distributions are given below. a.- The Morgenstern-Gumbel-Farlie Distribution (Johnson, N.L., Kotz, S. 1975) The joint survival and the first element of the vector of hazard function are given respectively by: GX) 'X 2 (x"x2) = Gx) (xd G x 2 (X2)[ 1+a.FX1 (xd Fx 2 (X2)] h X1 'X 2 (x"x2h = - lal < 1 ~l {LnG x ) (xI)+ LnG X2 (x2)+ Ln[l +a.Fx ) (xI)FX2 (X2)]) =hx a(I-G x2 (X2»)fX1 (xl) (Xl) - ------::;---"-) 1+a.Fx) (XI)FX2 (x2) where lal<l, and Xl,X2 >0. This distribution has the two special cases: (i).- When X}, X2 each have exponential distributions so that the corresponding hazards functions are constants, i.e., hj(xj)=hj j=I,2. Then the first component of the vector of bivariate hazard is: 41 h X I' x 2 (x}'x2h with J3 =[1-{J3exP(hlXl)-I}-l]hl = 1+[a{1-exp(-h2X2)-1}r 1 This particular distribution is a reasonable distribution if we consider that each of the marginal distributions is exponentially distributed and therefore has a constant hazard rate. The association between the components of the vector depends on the parameter a. We do not consider this distribution as the distribution of interest mainly because the parameter a does not have a direct interpretation. (ii).- When X 1, X2 each have Weibull distributions, then the marginal survival and the first component of the vector of hazard rates are respectively: G Xj (Xj) = exp ( -x~j ) Cj > 0, hx"x, (x]'x2h = [1- {pexp(x~,) where J3 = 1+[a{1- exp( Xj > 0; j = 1,2. Then -It}IXI -X~2 ) -1}]-l This distribution was not considered any longer because our maIO focus distributions with constant hazard rates. 42 IS on b.- Gumbel's Bivariate Exponential Distribution (Johnson, N.L., Kotz, S. 1975). The corresponding joint cumulative distribution, joint survival and the first component of the vector of bivariate hazard function are: Each marginal component has a hazard rate equal to one, and thus this distributional was not considered further. c.- Bivariate Exponential Distributions of Marshall and Olkin (Johnson, N.L., Kotz, S. 1975). The joint survival function has the form: G x l' X 2 (XJ, X2) = exp{-AIXI- A2 X2 -A12 max (Xlo x 2)} Al > 0, 1.. 2 > 0, 1.. 12 > 0; xl > 0, X2 > ° This distribution has the following nice properties such as: (i) its marginal distributions are exponentially distributed, so that their hazard are constant, (ii) it allows for different hazard rates for each of the components of the vector, (iii) the correlation between the 43 components is a function of 1. 12 , one of its parameters. The main reason for not using this bivariate distribution is that this is not an absolutely continuous bivariate distribution .. since the probability of the components being equal is greater than zero. We finally ended in using the bivariate exponential distribution of Sarkar because of its properties to be discussed in the next section. Ryu in 1993, extended the Marshall & Olkin's Bivariate exponential distribution such that it is absolutely continuous and need not to be memoryless. In addition, the new marginal distribution has an increasing hazard rate, and the joint distribution exhibits an aging pattern. This distribution offers an attractive alternative to the extention provided by Sarkar to the case of non constant hazard rates. In summary, several parametric bivariate failure time distributions, among these the family of the bivariate exponential distributions, were considered for application to this problem. Specifically, the Marshall-Olkin bivariate exponential distribution (Johnson & Kotz, 1975) was initially considered since its marginal distributions are univariate exponential, which is a common univariate distribution for the type of data we are interested in. However, it assigns a positive probability to the event of simultaneous occurrence of the failure of the two components of the vector, which is rarely the case in actual practice. In addition, this distribution also poses a probabilistic problem given the fact that it is not an absolutely continuous distribution, since it assigns a positive probability to the event {X=Y}. In many practical situations, like in rare diseases, an absolutely continuous bivariate exponential distribution is desirable. The bivariate exponential distribution proposed by Sarkar retains the property of exponential marginals distributions and it is absolutely continuous. 44 .. 3.3.2.- Definition and properties of the Sarkar's Bivariate Exponential Distribution: 3.3.2.1.- Definition. We will focus on the bivariate exponential distribution of Sarkar due to the properties this distribution has as going to showalter in this section. The random vector (X,Y) has the bivariate exponential distribution of Sarkar if the joint bivariate survival distribution is given by: P~X ~ x, Y. ~ y) = exp{-(A 2 - y A12 h}{I-[ A(Aly)r [ A(Alx)f+Y} *I[x<y] + y exp{-(A I - 1.. 12 )X}{l- [A(A2 x)r [A( A2Y)] I+Y} *I[x>y] where x>O, y>O, 1..1>0, 1..2>0, A12~O, y = '1 1.. 12 J\,I '1 + J\,2 and A(t)=l-exp(-t) for t>O. The joint bivariate density of (X,Y) is given by: fx,y(x,y) 1..1..* I )2 exp{-AI x - (1..2 + A12h} *{( Al + 1..2)( 1..2 + 1..\2) - 1..2Aexp( -AlY)} * Al +1.. 2 =( [A(A,Ix)r[ A(A,Iy)r(I+Y) *I[x<y] ( + A 1..* 2 )2 exp{-A,2Y-(A,1 +A,I2)X}*{(A I +A,J(A 2 +A,12)-A IAexp(-A,2 x)}* Al +1.. 2 [A( A,2Y)] Y[A( A,2 X)]-( I+Y) *I[x>y] 45 3.3.2.2.- Properties. ii) The marginal distributions of X and Yare exponential X - exp(A I + AIJ Y - exp( A2 + A12 ) ii) If A12=O, then X and Yare independent exponential distributions with parameters Al and A2 respectively. iii) The distribution of the minimum of X and Y is also exponential min(X, Y) - exp( A* = Al + A2 + A12 ) iv) min(X, Y) is independent of g(X,Y) for any g ill = {g( x, y): g(x, x) E ill where = 0; g( x, y) is strictly increasing (decreasing) in x(y) for fixed y(x) } v) The correlation between X and Y is given by: PX,Y AI2 AI2 ~ j! { j A( * )-1 j A( * )-I} =-* + -* L... * ( ). Ad 1 A + kA I + A211 A + kA 2 k=I k=I A A j=I A + Al + A J 2 From this last property we can see that when AI2=0 then px,y=O. Thus, AI2 can be interpreted as a measure of association between X and Y. The above properties make the Bivariate exponential distribution of Sarkar applicable to many practical situations such as eye trials or similar. In what follows we will assume that X and Y have the same marginal distribution, i.e. that AI=AI=A. This is a very reasonable assumption because it implies that the marginal failure time distributions are the same for both organs. 46 3.3.3.- Marginal MLE's of Aand y. Assuming that A1=A2=A, then y = ~~ ~ Al2 = 2Ay, and therefore: (a).- The joint bivariate exponential distribution of Sarkar is reduced to: fx,y(x,y) = A2 (1 + y) exp{-Ax - A(I + 2y)y}{1 + 2y - (I + y) exp(-Ay)}[A(AX)r [A(Ay)t1+Y\x<y) A2 (I + y) exp{-AY - A(I + 2y)x}{ 1+ 2y - (I + y) exp(-AX)}[A(AY)]Y [A(AX)]-(l+Y) Ilx>y) (b).- We can re-parameterize the marginal distributions of X and Y in terms ofS 1 as follows: x- exp(Sl = A(I + 2y)) Y - exp(Sl =A(I + 2y)) (c).- And the distribution ofthe minimum can be written in terms of S2 min(X, Y) - exp(S2 =2A(1 +y)) From (b) and (c) above, we can obtain the marginal MLE's ofS 1 and S2 as follows: - Since Sl can be estimated by either SI = X1 - or SI 1 = y; we average both estimators and - ="'1(1 2 X + Y1) we can estimate Sl by Sl - From the distribution of the minimum of X and Y we estimate S2 by S2 = X1 (1) Now, solving for the original parameters Aand y, we have: _ 1 I( 1 1) 1 1) 1 ( _ 1 X+y -~ !C r A = XII) - 2 X + Y , and y = 2 1 _ +1 X(1) 2 X Y 47 . 3.3.4.- Joint MLE's of A and y. We can rewrite the joint density defined in (a) above as: fX,Y (x,y) = fjJ,~ (x, y)I,x<Y) + fjZ~ (x, y)I,x,n ' and defining (; = {~ if X<Y if X > Y , • (1) ]O[ fx,Y(x,y) (2) ]1-0 5 = 0, . this becomes fx,Y(x,y) = [ fx,Y(x,y) The log likelihood for a single observation is given by: '- i = 10g(fx,Y (Xi, yJ) = 510g[ f*~~ (Xi, Yi)] + (1- 5) log[ f*:~ (Xi,y i)] Now, log[f*l,~ (x, y)] = 10g(1 + y) + 10g(A) + {-AX - A.( 1+ 2Y)Y} + log{ 1+ 2y - (1 + y)e- AY } + y 10g[A(Ax)] - (1 + y) log[A(AY)] and log[f*:~ (x, y)] = 10g(I + y) + 10g(A) + {-AY - A.( 1+ 2y)x} + log{ 1+ 2y - (1 + y)e- Ax } + ylog[A(AY)] - (1 + y) 10g[A(Ax)] Now, '-i can be rewritten as: 2 ,-.I = J j=l t 1 4 It. +5." 4 + (1-5.I )"1= 1£.J'l"IJ £.J~IJ.. , where \II .. j=l = 10g(1 + y) j=l t z =210g(A) 'IIi 1 =-AX i - A.( 1+ 2y)y i 'IIi2 = log{ 1+ 2y - (1 + y)e -AYi } 'IIi3 = ylog[ A(Axi)] 'IIi4 = -(1 +y) 10g[A(AyJ] ~l = -AY - A.( 1+ 2y)x i ~2 = 10g{I+2y-(I+y)e- AYi } ~3 = ylog[A(AYi)] ~4 = -(I+y)log[A(Ax i)] 48 In what follows we will denote by ~' = [A y]. To obtain the MLE's of A and y, we need to solve the estimating equations given by: Un(A) = tnat.a~ = 0 and . nat· Un{Y) = t ayl = 0, which are clearly non linear. Therefore the Newton-Raphson method is to be applied to iteratively obtain the MLE's. We will use as naive initial estimators those obtained as marginal MLE' s in the previous section. &t1 Defining e = ( A at = (at at) y), then ae aA' ay aAay I Eft ay2 J' then the iteration scheme to be used for each of the components of e is . 49 3.3.5.- Estimating Equations and Information Matrix In this section we develop the algebraic solution to the estimating equations and for the information matrix. 3.3.5.1.- Estimating Equations. By definition of f n of. n 2 i' the estimating equation for A. is given by a n 4 a n 4 a Un(A.)=~ o~ =~~OA. tj+~8i~OA. \IIij+~(l-8J~OA. ~ij where a OA. t l =0 a OA. t 2 2 = A. a Yi(l+y)e- AYi = 1+ 2y - (1+ Y)e-AYi OA. \IIi2 a (1+Y)Yie-AYi OA. \IIi4 = - a OA. ~2 [A(A.YJ] Xi(l+y)e- Axi = 1+2y-(1+y)e-Axj ~~ __ (1 +Y)Xie-Axi OA. 4 - [A(A.xJ] The corresponding estimating equation for Yis: . n Of. n 2 a n 4 a n 4 a j Un(Y) = ~ Oy. = ~~ Oy t + ~8i~ Oy \IIij + t(1-8i)~ Oy ~j where a -t Oy I 1 =-l+y a -t Oy 2 a =0 a 2-e- AYi = 1+ 2y - (1+ y)e-AYi Oy \IIil =-2A.Yi Oy \IIi2 ~ \IIi3 = 10~A(A.xJ] ~ \IIi4 =-log[A(AyJ] 50 .. o Cty 1;1 C 2 - e-)·x, 8y ~2 = 1+2y-(I+y)e-Ax, =-2AX i o ~ 1;3 = log[A(AyJ] 8y 1;4 = -lo~ A( AXJ] 3.3.5.2.- Observed Information Matrix. The observed information matrix is given by: iffl OACty I if f I, where each of the components of the matrix can be J 8y2 obtained as: &f . 02 n 4 if n 4 if (a).- :l'I2 = ~ :l'I2 = ~~ :1,\2 t j + ~8i~ :1,\2 \!Iij + ~(1-8J~ :1,\2 ~ij n iff. 1 1=1 VI\, VI\, n 2 1=1 .F1 VI\, 02 oA2 -t & aA2 \!IiI =0 & 1=1.F1 VI\, - 2 - - 2 oA yx~e-Axi [A(AxJ] (l+y)y~e-2AY, aA2 \!Ii4 = [1-e-Ay,t & if (l+y)y~e-AYi = - [1 +2y-(1 +y)e-AYi]2 - [1 + 2y -(1 +y)e- AYi ] yx~e-2Axi -2 1;1 = 0 01.: 2 (l+y)2y~e-2AYi \!Ii2 J=1 VI\, A2 if oA2 'Vi3 =-[I-e-Axit & - 1=1 (l+y)y~e-AY, + [A(AYi)) (l+y)2x~e-2Axi (l+y)x~e-Axi -2 ~i2 = 2- [ ] 01.: [1+2y-(I+y)e-Axi] 1+2y-(I+y)e-Axi yy~e-AYi [A(AYJ] 51 .. 82 fJy2 1 't 1 if fJy2 \IIil if fJy2 \IIi3 if fJy2 ~1 if fJy2 ~3 82 =0 =- (1 +y)2 fJy2 =0 -2 \IIi2 =2 fJy [1+2y-(1+y)e- AYi ] =0 fJy2 \IIi4 't 2 if if (2_e- AYi )2 =0 =0 =0 if fJy8'A. 't 1 = 0 if fJy8'A. \IIil = -2Yi if (1+y)(2-e- AYi hie-AYi Yie-AYi fJy8'A. \IIi2 =- [1+2y-(1+y)e-AYit + 1+2y-(1+y)e-AYi if fJyOA. \IIi3 = xie- Axi [A( 'A.xJ] 82 fJy8'A. \IIi4 = 52 Yie-AYi [A( 'A.y J] & fJy8'A ~il 82 =- 2x j (1+y)(2-e- Ax ')x j e- Axi xje- h , fJy8'A ~i2 =- [1+2y_(1+y)e-Axi]2 + 1+2y-(1+y)e- h & fJy8'A ~3 Yje- AYi = [A('AYJ] ;j fJy8'A ~j4 x·e- h , , =- [~('AxJ] Since the components of the observed information matrix are expressed as a sum of independent random variables, by Khintchine's WLLN, the observed information matrix converges to the expected information matrix (Sen & Singer, 1994, pp 206). " 53 CHAPTER 4 .. Model for Censored Bivariate Exponential Distribution of Sarkar We now consider the more realistic case when the failure times of any of the two organs are censored. 4.1.- Censoring schemes Let C be the censored random variable which we will assume IS exponentially distributed with parameter I..l, and independent of(X,Y). Censoring can happen in three different ways: i.- Total Censoring: Censoring comes before the smallest failure time, i.e. C < ~l) ii.- Partial Censoring: The censored time is in between the smallest and the largest failure time, i.e. ~l) < C < ~2) iii.- No censoring at all: the smallest and the largest failure time are observed, C>~2). The following two indicator variables are defined in order to identify the type of censoring corresponding to a particular observation: t, t, ={~ if C <X(l) if C > X(l) ={~ if X(I) < C < X(2) otherwise 54 • Thus, for total censoring (11 ,12)=(1, 0), for partial censoring (11 ,12)=(0, 1), and for no censoring (11 ,12)=(0, 0). • 4.2.- Model Specification and Maximum Likelihood Estimation. When censoring is total, we only observe the censored time and therefore the only available information we have can be represented by the observed event E 1={ C=c ; ~1»c}. When censoring is partial we observe the minimum failure time and the censored time, the available information can be represented by the observed event E2={~1) ~I) ; C=c and ~2»C}. Similarly for the uncensored case E3={~1)~1), ~2)~2); C>X(2) }. Therefore, the i-th observation, i=l, 2, .... n, can be represented as 1t~Ii1t~2i1t~-tli-t2i, where 1tj is defined as follows 0=1,2,3), for all i=l, 2, ....,n . 1t 1=fc(Ci)*P{ Xi(1»cd, where fc is the density function of the censored time which is known has exponential distribution with parameter fJ., and 1t 2 = r f xW ,x(2) (xi( I)' Y) *fc {cJdY , and Ci The likelihood for a sample of size n therefore can be expressed as: 55 Alternatively, we can write the likelihood as: L: n where !fC<c;)' Fx,,, (cJ!", {I fxen,x", (X;iJ) ,y) •fc{c;)dY}'" {fX(Ij,x(2) (Xi( J) ,X i(2)) * Fe (Xi(2))} Fx (Ij and Fe !-tjJ-tj2 represent the survival function of ~!) and the censored variable respectively. Evaluating the likelihood piece by piece, we obtain: (a).- f e (c.) 1 = lie ,.... -l!Cj . , FX(I) (c.) = e -21..( 1 !+Y)Cj However, by the definition of the joint density we have • (s=1,2) can be expressed as follows: j f~~~ (xi(I) , y)dy = (1 + y)[ A( AXi(I)) Cj r Ae -J.xj(J) * After consecutive integration by parts of the integral on the right side above, we have that 56 which can be expresed as e-2A.YCj [A(ACJr times a power series on A(ACi), called S(A(ACi)). Therefore, 1Ci f~~~ (Xi(l) ,y)dy =(I + Y)[A( AXi(l))r Ae-A.xi(l) e-2A.YCi[A(ACi)r s( A(ACJ) =(I + Y)[A( AXi(l) )A(ACi)r Ae -~xj(1l+2ycd~ A(AC j)) Finally, the last component of the likelihood is given by: . Therefore, the likelihood can be expressed as: n r {~e-llCje-2A(1+Y)Ci i1 {(I + Y)[A( AXi(l) )A(A.cJr A.e -~Xj(l)+2YC;)S(A(ACJ) fi2 L= {{ f~~~ (Xi' Y J} {f~~~ (Xi' Yi)} i 6 j H t e-~(2) } l-til- i2 The loglikelihood is thus given by: n f = L 'til {log~- ci(~+2A.(1 +y))} + i=l ~ 't i2 IOg{( 1+Y)[A( Axi(l) )A(ACJr Ae-~Xi(I)+2YC;)S(A(ACJ)} + t(1=1 1- 'til - 'ti2){Oi logf~~~ (Xi' Yi) +( I-Oj) logf~~~ (Xi, Yj) - ~Xj(2)} 57 , where ~il = 10gJ..l ~i3 = -J..lX H2 ) U1 = 10g{A(I+y)} U2 U4 = 10g{S(A(AcJ)} \IIj2 \IIi3 = ylog[A(AxJ] \IIi4 = yIOg{A(Axj(I))A(AcJ} = log{ 1+ 2y - (1 + y)e- AYi } =-(I+y)log[A(AYi)] 1;1 =-AY-Ml+2y)x i 1;2 = log{ 1+ 2y - (1 + y)e -AYi } 1;3 = ylog[A(AyJ] 1;4 = -(I+y)log[A(AxJ] In what follows we will denote by ~ = [A y J..ll To obtain the MLE's of A, y and J..l, we need to solve the estimating equations given by: clearly non linear, therefore the Newton-Raphson method is to be applied to obtain the MLE's. As in the uncensored case, we will use as naive initial estimators, those obtained as marginal MLE's in the uncensored case for A and y, and for J..l its corresponding marginal MLE . 58 I ele . Defining Oe (oe Oe oe) 0).' (}y , Oil ; oe = I 0).2 I &e &e oeoe' =1 (}yo). - - I &e lOllo). ele l I o).(}y o).ollj a2 e &e I fie (}y2 02e 01l(}Y (}yOIl I &e I 0112 J and the iteration scheme to be used is the same described in chapter 3. • 59 4.3.-Estimating Equations and Information Matrix. 4.3.1.- Estimating Equations. The estimating equation for A is given by: nat. n 2a 4a n Un(A) = ~ a'll = ~til~a'l ~jj + Ltj2~a'l Ujj + 1=1 I\, 1=1 .F1 I\, 1=1 )=1 I\, t( 1- til - t i2 )Oj ±a~ \j!jj + I( 1- til - t i2 ){(1-0J± a~ ~jj +~i3} 1=1 .F1 1=1 .F1 I\, I\, where a aA ~il = 0 a aA ~i3 =0 a 1 aA Uil = A a • c ie -A.ci A( AX j(1)) + Xi( 1) e -AXj(I) A{ ACj) aA Uj2 = A( AxjO) )A( ACJ : Uj3 = -(Xi(I) +2yC i ) : Ui4 = : logS(A(AC i )) a aA \j!i2 a aA \j!j3 a yxje- AXi = [A{ AXJ] aA 1;1 =-Yi -(1+2y)x j Yi(l+y)e- AYi = 1+2y-(1+y)e-AYi a aA \j!i4 = - (1 + y)Yje- AYi [A{ AyJ] a Xi(l+y)e- Axi aA 1;2 = 1+2y -(1 +y)e- Axi Ax ~I; __ (1 +y)xie- ; aA 4 [A{AxJ] 60 . where a a r 02 = -2ACoI Or~l Or l;il = 0 a Or l;i3 =0 a 1 -u - - Or il - 1+y a =-2ACo Oru13 0 1 a 2 - e-A.Yi Or \IIi2 = 1+2y-(1+y)e-A.Yi ~ \IIi3 = log[A{AxJ] Or \IIi4 =-log[A{Ay J] a Or ~l =-2AX i a Or ~3 = log[A{AyJ] • a Or \IIil =- 2Ay i a Or ~2 = 1+2y-(1+y)e-A.xi a Or ~4 =-lo~ A{MJ] where a alll;il a alll;i3 1 = 11 =-Xj(2) 61 o -u··=o Oil IJ o -~ oil j=l, 2, 3, 4 .IJ =o j=1,2,3,4 j=1,2,3,4 4.3.2.- Observed Information Matrix. 02 £ o'A.Oy &£ Oy2 02 £ °IlOy &£ l -I O'A.OIl &£ I OyOIl I ,where 1 &£ Oll2 1 J (a).- , and & 0'A.2 l;ij =0 j=l, 2, 3 0'A. & il - 0'A.2 \j!ij 'A.2 2 -Ac ::12 2 0 1 -u 2 --- - u o'A.2 . u - i2 - Ci [( e 1 A 'A.c i )] 2 x 2 e-Ax;(l) i(I) - -----'-----'--:[( A 'A.xi( 1) )] 2 02 and o'A.2 ~ij are the same as those derived in chapter 3 for j= 1, 2, 3,4. 62 (b).- 82 ()y2 ~ij =0 j=l, 2, 3 EP ()y2 if ()y2 Ui3 if ()y2 EP 1 U il - 2 Ui2 =- {1+y)2 ()y 02 =0 ()y2 Ui4 =0 if = ()y2 logS(A(AcJ) if \j!ij and ()y2 ~ij are the same as those derived in chapter 3 for j=l, 2, 3,4. and 02 if 0J.!2 Ujj = 0 , 0J.!2 02 \j!ij = 0 and 0J.!2 ~ij = 0 for j=l, 2, 3, 4. 63 02 (}yOA ~il 02 (}yOA Uil ~~ \j!ij 82 (}yOA ~i2 =0 82 =0 and (}yOA U i2 ~~A~ij = -2c i 82 (}yOA ~i3 82 =0 (}yOA Ui3 =0 = -2c i are the same as those derived in chapter 3. 02 ~OA ~ij = 0 for j=l, 2, 3. & ~aA Uij 02 o~(}y ~ij & 0J.U7y Uij 02 =0 , o~oA \j!ij =0 for j=l, 2, 3. 02 = 0 , o~(}y \j!ij = 0, and • & o~A ~j =0 02 = 0 , and a~(}y ~ij 64 for j=l, 2, 3, 4. . =0 for j=l, 2, 3,4. Large sample properties for the score statistics can be applied to make inferences about the parameters involved in both the uncensored and the censored case. • 65 CHAPTER 5 Group Sequential Test 5.1.- Introduction In clinical trials in which more than one observation is obtained from the same individual, such as ophtalmological studies where both eyes are analyzed, or when two meaningful endpoints are measured, group sequential interim analysis methodology must be applied. The methodology developed in chapters 3 & 4 of this dissertation for the bivariate exponential distribution of Sarkar is used to develop group sequential interim analysis method. Our model incorporates information from each of the individual organs, which is more efficient in the sense that marginal analysis can also be performed in the traditional way. Group sequential test statistics based on the bivariate exponential distribution of Sarkar are presented under the absence and presence of censoring. In what follows, we first present the setup for repeated significance test statistics and we show that the discrete sequence process generated by the application of repeated test along the time converges to a Brownian motion process. We use the above convergency property to obtain group sequential boundaries for the bivariate exponential distribution of Sarkar. • 5.2.- Bivariate Group Sequential Test Let us consider a clinical trial where subjects are randomly allocated into • either an experimental or control treatment. Let us assume further that an unspecified number of interim analyses, K, are planned to be done during the study period, and that the decision of stooping the trial is based on repeated significance test statistics after evaluation of each group. Let t denote the study time and that the interim analyses are performed at the point times tl, h, ....., tK, where K is unspecified in advance, and 0<t1< t2< .....< tK. To make statistical inferences under the Group Sequential framework, it is necessary to show that the discrete sequence of Score statistics at the different time points converges to a Brownian Motion Process At a particular monitoring time tk (k=l, 2, ....., K) we have 5 possible situations: {X < Y ~ tk} Both X and Y failed before tk {Y<X~td {X < tk < Y} One of then failed before tk {Y <tk <X} {X> tk ; Y> td None of them failed before tk Therefore for tk, the likelihood, for a single observation, is given by: ~~, tk T ( ) (j) Jk = 5 [ gx,y(x,y) E j , where • Ok1=I{X < Y ~ td (1) ( gx,y x,y) (1) ( = fx,y x,y ) Ok2=I{Y < X ~ td (2) ( gx,y x,y) (2) ( = fx,Y x,y) 67 (3) ( gx,y x,y ) = OC!f (1) ( fX,y x,y ) dy lk (4) ( gx,y x,y ) = OC!f (2) ( fX,y x,y) dx lk Define the Score statistics and the information at time point t respectively We have that E[U n (8;t)] that Ie(t = 0) = ° = 0, and that Ie(t) is an increasing function in t, such and Ie(t = 00) = Ie • By the Central Limit Theorem, for each tk we have that: lun(e;tk)=wn(e;t k) ~ N(O,Ie(tk)) Let us now consider the time points tl and h (tl < h), and denote .3(tl) the set of all possible events generated up to the time tl -----1------1 L(tl) h L(h) h The conditional likelihood up to the time h given the history up to the time h is L(h 1.3(tl» and it is represented by: 68 1 if 8 11 =1 • 8 12=1 f y(t 2)[ (2) r-Iy(t gx,YI013=1 h X,Ylo13=l [ (l) [ (I) gX,YI0 14 =1 or f X(t 2)[ (2) r-IX(t hX,Ylo I 4=1 Ell5 [gX,YlOls=l t2r (r) if 8 13=1 if 8 14=1 if 815=1 where ly(h)=I{Y < h} and Ix(h)=I{X < h} and 02r are defined as above The expressions for the components of the likelihood are simply given by: (1) ( (1) fx,Y x,y) . gX,YlolS =1 = k.... .I.X,Y (t b t)1 r<i~{x,y) (2) gX,YI81s=1 = k.... .I.x,Y (t b t)1 00 -f f XY (2) ( x,y) dx (4) t2' gX,YI81s=1 = --=-F--"(-t-t-:-)X,Y b 1 69 r \ Now, L(e; h)=L(e; tl)* L(t2 3(tl)) , then 1 log L(e; h)=logL(e; tl)+log L(h I 3(tl)) log L(h I 3(tl))= log L(e; t2)-logL(e; tl) Thus, Un ( t 213{ t 1)) = Un ( t 2) - Un ( t 1) Similarly for any pair tk-I and tk, the increments in the scores are [U n( t 113( to)); Un ( t 21 3(t 1)); ; Un ( t K 13{ t K-I))]. The vector of components has independent elements (conditionally independent). The sequence of conditional scores {U n( t k 13( t k-1)): k = 1,2,.... K} is a zero mean martingale and it converges to a Brownian Motion Process (Weak convergence of martingales) As a consequence of the result above, we can use the Lan & DeMets spending function to obtain the boundary crossing probabilities for the developed model. 70 5.3.- Concluding Remarks The proposed model is based on the Bivariate Exponential Distribution of .. Sarkar. This bivariate distribution is absolutely continuous with marginal hazard rates being constant and non-negatively correlated. These properties make this distribution the appropriate choice for several real life problems such as eye studies or any type of study comprised of pairs of organs. The proposed method allows one to use the information coming from each of the organs being analized, and therefore it allows for no loss of information such as when one uses a summary measure of the participating organs. In a particular clinical trial application when the outcome of the two organs can be assumed to follow the bivariate exponential distribution af Sarkar, one may utilize the proposed method to obtain estimates of the parameter A. accounting for the correlation between organs. Testing of the hypothesis for treatment differences can also easily be constructed based on the A.' s of each treatment group. Since the test statistic can be calculated at any given tieme point, and the discrete process generated converges to a Brownian motion, one can utilize the Group Sequential boundaries for repeated testing on accumulated data. 71 CHAPTER 6 Numerical Results 6.1.- Introduction We have chosen to simulate the behavior of the methodology in a given hypothetical sample generated from the bivariate exponential distribution of Sarkar. Thus, we are not examining the goodness of fit of the parametric model, but rather the behavior of the procedure in a situation where it applies. In real applications, one could examine the goodness of fit of the bivariate exponential distribution of Sarkar by using its properties, i.e., examine if the minimum failure time of the two organs distribution has a constant hazard. We compare the methodology to the standard method of using the rmmmum failure time of the two organs, under a variety of scenarios of different correlations. These are outlined in table 6.1. First, we simulate 500 observations from the bivariate exponential distribution of Sarkar for both an treatment and a control group. The simulated distributions are such that the corresponding mean response time for each of the participating organs is 20 months for the experimental group and 16 for the control group. This represents a • treatment difference in the survival of 4 months. between the two organs are also considered. Different degrees of correlation Table 6.1 contains (a) the different scenarios (true parameter values) used for simulations and (b) descriptive statistics of the simulated sample. . 73 Table 6.1. Simulation Scenarios and values for sample of size 500: Bivariate Exponential Distribution of Sarkar Experimental Group Control Group n=500,E[X]=E[Y]=16 n=500,E[X]=E[Y]=20 True values A Sample values True values Al =0.04 x=19.8 Al =0.0500 x=16.3 AI2 =0.01 Y=20.5 AI2 =0.0125 Y=14.5 r=0.19 B r=0.13 Al =0.025 x=19.8 Al =0.0300 X=15.6 AI2 =0.025 Y=19.7 A12 =0.0325 Y=14.6 r=0.45 C Sample values r=0.49 Al =0.01 x=20.9 Al =0.0200 x=16.8 A12 =0.04 Y=20.0 A12 =0.0425 Y=15.9 r=0.82 r=0.65 74 ~ As can be seen in figure 6.1 (two-way scatterplots and univariate box plot), the marginal distributions of the simulated data are similar under two different scenarios with different correlations. In addition, the bottom set of graphs with higher correlations display a tighter cloud for the bivariate distribution, as expected. 75 Figure 6.1. Simulated Bivariate Exponential Distribution of Sarkar Simulateo SSr'1r8r'S BI\lI!lr'l8tE- rlponl?ntldJ n=500, mean=20 C'lstrlCu!ton SllllulBIe>a Sar ... ars, CQ/"r:O.17 n=500, llean=16 l -=----- 8l~arlalE' • • !, 200 200 0 lSO '50 e:.oonpnoal • D1StrlOlJt~on cOI'"I"':0.l7 '!I' II J 0 0 '00 50 0 ~ 9rS't9 lOa 0 0 ~ R£ct> 50 lOa v S~lIlulatec Sar-kar's Billilr1ate [):ponentlal n=':500. IIIrlln:20 COI""'''Q 75 150 200 ~ 50 0 0 k 00 <5Z> ® 50 tOO v S1111ulateo Sarkar'g Bhartate E):DonenUal D~stntlul1on ,>0 200 ~ O~slrltJuUon n=500. IIrl!!f'l=J6. co",."O.68 200 '00 --=- 0 ISO '00 50 0 J; lSO 0 1~ocB 50 _otto 0 laO 0 000 '00 v lSO 200 ~ 76 50 0 00 fPJO~ 50 0 ,00 y '50 200 ~ 6.2.- Single Analysis Results at the end of the Study. .. 6.2.1.- Ignoring the bivariate nature of the data. As mentioned in chapter 2, most analyses of bivariate failure time data is reduced to a univariate situation such as by studying the time to failure of the first organ. In our setup, this is essentially studying the time of the random variable min(X,V). bivariate exponential distribution of Sarkar min(X,Y) - exp(2A.+A.IZ). In the Figure 6.2 displays the Kaplan-Meier estimates of the survival curves for the dataset for the experimental group and the control group under scenario A. summary statistics for all three simulation scenarios. 77 Table 6.2 displays Figure 6.2. Kaplan-Meier Survival Curves for Scenario A Graue> 0 .. Graue: J 1.00 - . >- 0.75 - D fU D o !l 0.50 - fU > > L :J U1 0.25 - 0.00 I o I I 20 10 I 30 I ~o I 50 tlme Kaplan-Meler Survlval Curve logrank test: '1. 2 = 21.33 p<O.OOOI • 78 • Scenario . . fior Scenanos A , B ande T abl e 62 .. Summary Statlstlcs 2 Events Median Group Obs Mean X Control 500 429 6.69 8.80 A Experimental 500 406 7.97 12.09 Control 500 412 7.12 10.11 B Experimental 500 406 9.68 13.15 Control 500 402 9.25 9.25 C Experimental 500 387 12.46 . • 79 p 12.46 21.33 0.0000 14.10 0.0002 20.07 0.0000 6.2.2. Incorporating the Bivariate Nature of the Data. • In this section section we present the analysis results comparing the experimental to the control group with no interim analyses and complete data. We now consider the test procedure using the bivariate information under the four scenarios presented in table 6.1. We use the score statistics to compare AE to Ac , i.e., we compare the distributions treating y as a nuisance parameter. A program in e++ was written to obtain the estimators and standard errors by using the Newton-Raphson procedure described earlier. To test Ho: AE = Ac we will use the statistic T= In( Ln( ~E) - Ln( ~c)) - N( 0, ~) or the equivalent Z =.J2rl( Ln( ~E) - Ln( ~c)) - N( 0,1). The results for the three different scenarios are shown in table 6.3. 80 Table 6.3. Test Statistics for the Bivariate Exponential Distribution of Sarkar Scenario Estimated Parameters Z Statistic p-value Z=-11.38 P = 0.0000 Z = -7.93 P = 0.0000 Z = -26.14 P = 0.0000 A A AE = 0.0360 Ac =0.0516 A B AE = 0.0249 Ac = 0.0320 A C AE = 0.0091 A Ac = 0.0208 81 6.3.- Group Sequential Boundaries .. Interim analyses on the simulated censored bivariate distribution of Sarkar, are based on the Z statistic Z = .J2r1( Ln( ~E) analysis on the accumulated data. - Ln( ~c)) - N( 0,1) ,.computed at each interim The corresponding p-value for the k-th interim analysis is computed by using as the accumulated level of significance a..Jt:, where tk is the proportion of elapsed time at to the point tk. Our numerical results are based on the assumption of a total study time of 60 months, and interim analyses are going to be performed at 12, 24, 36 and 48 months. Therefore tl=12/60=0.2, t2=0.4, t3=0.6, t4=0.8 and t5=1. By using a program written in e++, the estimated parameters and Z statistics A for the interim analysis at time point 12 months, for the scenario A, are respectively: AE A = 0.035, Ac = 0.052, Z=-12.52. The level of a..Jt: =0.05,J0.2=0.0224, significance of 5% gIves a rejection level of which implies a rejection of the null hypothesis of no difference between the lambdas. Therefore an early termination of the trial must be considered because of the observed treatment difference. 82 CHAPTER 7 .. Future Research The methodology presented in this research allows the investigator to use the bivariate nature of the data and therefore it is not necessary to restrict the analysis to a single result summarizing the pair of outcomes by using either the minimum of the failure of the organs or some other summary statistics. In building the group sequential boundaries for censored bivariate survival data, we have considered a series of parametric bivariate models with different functional forms and underlying assumptions. Having in mind the applicability of these parametric distributions in real life data, we focused our interest in a kind of model that fits in situations in which both marginal distributions have constant and equal hazard rates such as the family of bivariate exponential distributions. After discussing the applicability of several bivariate exponential distributions, we used the bivariate exponential distribution of Sarkar since it is an absolutely continuous random vector and fulfills both the mathematical and the applicability requirements. We restricted_ the Sarkar's exponential bivariate distribution to the situation in which both marginal distributions have constant hazard rates. This constraint was done in order to satisfy many real life bivariate 83 survival data, such as ophthalmological data where illnesses with constant hazard rates in each eye is a reasonable assumption. .. However, there are a series of situations in which this assumption may not work and different hazard rates must be considered. As an example of this, let us consider a study where our main outcome variable is the "survival" in breast cancer in women. Kelsey & Hom-Ross (1993) reported a study on breast cancer in which there is an excess of left-sided tumors, with ratios in the range of 1.05 to 1.20. Therefore, in a clinical trial on breast cancer prevention, it would be more nearly correct to assume a higher hazard rate for the left breast than for the right breast in the placebo group. Another study on testicular cancer by Stone, Cruickshank, Sanderman & Matthews .. (1991), reported that in men with unilateral cryptorchidism, there would be a much more higher rate of testicular cancer on the cryptorchid side, even though rates on both sides are above the rates in normal men. The above examples show that a bivariate model that incorporates different hazard rates has actual applications in cancer epidemiology. Therefore, future research must include the study of different hazard rates for each of the individual organs of the patients. This assumption of non equal hazard rates allows the investigator to compare pairs of organs at different stages of the disease and therefore with different hazard rates. The bivariate exponential distribution of Sarkar allows unequal but constant hazard rates in each of the organs (AI 84 :t. A2). This extension . , follows the same line of the development presented here and it adds some extra terms to the equations used in the derivations and therefore it can be easily done. .. We have assumed that data fit the bivariate exponential distribution of Sarkar. However, departures from this assumption may invalidate the results we have obtained. A possible extension of this research would be to investigate the robustness of the results. One could examine the behavior of the test statistics for data with non-constant marginal hazards. A non-constancy parameter could be introduced and its effects studied with further simulations. Another extension of the methodology presented here is to use a bivariate • parametric distribution which allows non constant hazard rates, such as the bivariate exponential Weibull distribution. This distribution can be applied also to situations in which both the non constant and the non equal hazard rates of the participating organs is the most realistic situation. Additional areas of research to be considered are sample-size estimation and interval estimation after group sequential test. To develop these statistical areas would improve the group sequential design and consequently their use in clinical trials when bivariate survival data are generated. 85 Other adjustments to the methodology presented here are related to studies when more than two observations are generated on the same individual such as dental studies, or in studies when clusters of size bigger than two are the unit of analysis such as in family or community studies. " . 86 References: Beny, D.A (1985). Interim analysis in clinical trials: Classical vs. bayesian approaches. Statistics in medicine vol 4, 521-526. Beny, D.A (1989). Monitoring accumulating data in a clinical trial. Biometrics 4~, 11971211. Cox. D.R (1972). Regression models and life-tables (with discussion). J.R. Statist. Soc. B 34, 187-220. Cox, D.R, Oakes, D. (1984). Analysis ofsurvival data. Chapman and Hall, New York. DeMets, D. and Gail, M.H. (1985). Use ofLogrank tests and group sequential methods at fixed calendar times. Biometrics 41, 1039-1044. • .. DeMets, D.L. (1987). Practical aspects in data monitoring: A brief review. Statistics in Medicine vol 6, 753-760. DeRouen, T.A, Manel, L. Hujoel, P. (1991). Measurement of associations in periodontal diseases using statistical methods for dependent data.. Journal ofPeriodontal Research 26; 218-229. Donald, A, Donner, A (1987). Adjustment to the Mantel-Haenszel chi-square statistic and odds ratio variance estimator when data are clustered. Statistics in Medicine 6; 491- 499. Donner, A (1989). Statistical methods in Ophtalmology: An adjusted Chi-Square approach. Biometrics 45,605-611. Donner, A, Brown K.S., Brasher, P. (1990). A methodological review of non-therapeutic intervention trials employing cluster randomization, 1979-1989. International Journal ofEpidemiology, vol 19, #4: 795-799. • . Fleming, T.R, DeMets, D.L. (1993). Monitoring of Clinical Trials: Issues and Recomendations. Controlled Clinical Trials 14, 183-197. Freedman, L.S., Spiegelhalter, D.l (1989). Comparison of bayesian with group sequential methods for monitoring clinical trials. Controlled clinical trials 10,357-367. Haybittle, J.L. (1971). Repeated assessment of results in clinical trials of cancer treatment. British J. Radiol. 44, 793-797. Helms, R (1992). The joy ofmodeling: General linear models for the analysis of continuous data. Copyrighted unpublished manuscript, The University of North Carolina, Chapel Hill, NC 27599-7400. • Hujoel, P.P., Weyant, RJ., DeRouen, T.A. (1991). Measures of dental disease occurrence. Community Dental Oral Epidemiology 19; 252-256. Jennison, C., Turnbull, B.W. (1990). Statistical approaches to interim monitoring of medical trials: A review and commentary. Statistical Science vol 3, 299-317. Jennison, C. and Turnbull B. (1991). Group sequential tets and repeated confidence intervals. Handbook ofSequential Analysis (Ghosh, B.K and Sen, P.K eds) 12,283-311, Marcel Dekker Inc, New York. Johnson, N.L., Kotz, S. (1975). A vector multivariate hazard rate. Journal ofMultivariate Analysis 5, 53-66 Kelsey, J.L., Hom-Ross, P.L. (1993). Breast Cancer: magnitude of the problem and descriptive epidemiology. Epidemiology Reviews: 15: 7-16. Kupper, L.L., Haseman, J.K (1978). The use of a correlated Binomial model for the analysis of certain toxicological experiments. Biometrics 34, 69-76. Lan, KKG., DeMets, D.L. (1983). Discrete sequential boundaries for clinical trials. Biometrika 70, 659-663. Lan, KKG. and Lachin, J.M. (1990). Implementation of group sequentiallogrank tets in a maximum duration trial. Biometrics 46, 759-770. Lan, KKG., Reboussin, D.M., DeMets, D.L. (1993). Information and information fractions for design and sequential monitoring of clinical trials. To appear in Comm. in Stat. seriesA. Lee, E.W., Wei, LJ., Amato, D.A. (1992). Cox-type regression analysis for large numbers of small groups of correlated failure time observations. J.P. Klein and P.K. Goel (eds.), Survival Analysis: State ofthe art, 237-247. Kluwer Academic Publishers. Printed in Netherlands. Liang, KY., Zeger S.L. (1986). Longitudinal data analysis using generalized 'linear models. Biometrika 73; 13-22. 88 . Mantel, N. (1966). Evaluation of survival data and two new rank order statistics arising in its consideration. Cancer Chemo. Rep. 50, 163-170. . O'Brien, P.C., Fleming, T.H. (1979). A multiple testing procedure for clinical trials. Biometrics 35, 549-556. Peto, R. (1972). Rank test of maximal power against Lehmann-type alternatives. Biometrika 59,472-475. Peto, R., Pike,M.C., Armitage, P. et al (1976). Design and analysis of randomized clinical trials requiring prolonged observations of each patient: 1. Introduction and design. British 1. Cancer 43, 153-162. Rosner, B. (1984). Multivariate methods in Ophtalmology with application to other paireddata situations. Biometrics 40, 1025-1035. Ryu, Keunkwan. (1993). An extension of Marshall and Olkin's Bivariate Exponential Distribution. Journal o/the American Statistical Association vol 88, number 424, 14581465. • Pocock, SJ. (1977). Group sequential methods in the design and analysis of clinical trilas. Biometrika 64, 191-199. • Sarkar, S.K. (1987). A Continuous Bivariate Exponential Distribution. Journal of The American Statistical Association, vol 82: 398, 667-675. Sen, P.K., Singer, lM. (1993). Large Sample Methods in Statistics: An Introduction with Applications. Chapman & Hall, New York. Sorbinil Retinopathy Trial Research Group (1990). A Randomized Trial of Sorbinil, an aldose Reductase Inhibitor, in Diabetic Retinopathy. Archives ofOphtalmology, vol 108, 1234-1244. Sterne, J.AC., Johnson, N.W., Wilton, J.M.A, Joyston-Bechal S, Smales F.C. (1988). Variance components analysis of data from periodontal research. 1. Periodontal research 23; 148-153 . • Stone J.M., Cruickshank, D.G., Sanderman T.F., Matthews, J.P. (1991). Laterality, maldescent, trauma and other clinical factors in the epidemiology of testis cancer in Victoria, Australia. British Journal ofCancer: 64:'132-13 8 Tsiatis, AA (1981). The Asymptotic Joint Distribution of the Efficient Score Test for the 89 Proportional Hazard Model Calculated over Time. Biometrika 68: 1, 3 11-3 15. Zeger, S.L, Liang, K. Y. (1986). Longitudinal data analysis for discrete and continuous outcomes. Biometrics 42; 121-130. 90 "