Bayesian Assurance and Sample Size Determination for Experimental Studies
Skip to main content
eScholarship
Open Access Publications from the University of California

UCLA

UCLA Electronic Theses and Dissertations bannerUCLA

Bayesian Assurance and Sample Size Determination for Experimental Studies

Abstract

Determining the sample size to meet the inferential objectives of a study is of central importance in experimental design. There is an extensive collection of methods addressing this problem from diverse perspectives. The Bayesian paradigm, in particular, has attracted noticeable attention and includes different perspectives for sample size determination. While traditional Bayesian methods formulate sample size determination as a decision problem that optimizes a given utility functions (Lindley, 1997), practical experimental settings may require a more flexible approach based upon simulating analysis and design objectives (see, e.g., O'Hagan and Stevens, 2001). Building upon the latter approach, we devise a general Bayesian framework for simulation-based sample size determination using Bayesian assurance that can be easily implemented on modest computing architectures. We qualify the need for different priors for the design and analysis stage, working primarily in the context of conjugate Bayesian linear regression models, where we consider known and unknown variances. We also compare the assurance to a utility-based approach that involves the specification of objective functions to determine the rate of correct classification (Inoue, Berry, and Parmigiani, 2005). Throughout, we draw parallels with frequentist solutions, which arise as special cases, and alternate Bayesian approaches with an emphasis on how the numerical results from existing methods arise as special cases in our framework.

We further extend our conjugate linear model's capabilities to encompass the multiple testing framework, where the assurance is now characterized by conditions placed on the Bayesian false discovery rate (FDR). Under this framework, we investigate the effects of multiple comparison adjustments on assurance and sample size determination. Adjustments include enforcing different assigned threshold values for the Bayesian FDR and conditions related to the credible interval condition, and varying the number of pairwise hypothesis tests being conducted. Of particular interest is observing how the number of pairwise tests being conducted affects the assurance under fixed constraints placed on the Bayesian FDR as defined in Muller et al., 2004. We assess how our proposed model performs in commonplace large-scale problems, specifically microarray data. Our methodology is implemented in a study of mammary cancer in the rat, where four distinct patterns of expression are provided. Future tasks involve assessing how our method performs when comparing more than two subgroups and enforcing objective ways of choosing optimal threshold values.

This dissertation captures the vast applicability of the two-stage framework, offering a robust Bayesian approach for sample size determination equipped for addressing a wide selection of problems taking place both within and outside clinical trial settings. There is broad potential for growth and development in the methods introduced, with numerous routes available for future exploration.

Main Content
For improved accessibility of PDF content, download the file to your device.
Current View