[talks] Yingfei Wang will present her FPO, "Advances in decision-making under uncertainty: inference, finite-time analysis, and health applications" on Tuesday, 5/9/2017 at 1pm in CS 302.

Nicki Gotsis ngotsis at CS.Princeton.EDU
Thu Apr 27 15:01:06 EDT 2017


Yingfei Wang will present her FPO, "Advances in decision-making under uncertainty: inference, finite-time analysis, and health applications" on Tuesday, 5/9/2017 at 1pm in CS 302. 

The members of her committee are: 
Adviser: Warren Powell (ORFE) 
Readers: Bernard Chazelle and Mengdi Wang (ORFE) 
Examiners: Szymon Rusinkiewicz, Han Liu (ORFE) and Warren Powell (ORFE) 

A copy of her thesis is available in Room 310. Everyone is invited to attend her talk. The talk abstract follow below: 

This thesis considers the problem of sequentially making decisions under uncertainty, 
exploring the ways where efficient information collection influences and improves 
decision-making strategies. Most previous optimal learning approaches are restricted 
to fully sequential settings with Gaussian noise models where exact analytic solutions 
can be easily obtained. In this thesis, we bridge the gap between statistics, machine 
learning and optimal learning by providing a comprehensive set of techniques 
that span from designing appropriate stochastic models to describe the uncertain environment, 
to proposing novel statistical models and inferences, to finite-time and 
asymptotic guarantees, with an emphasis on how ecient 
information collection can expand access, decrease costs and improve quality in health care. 
Specifically, we provide the first finite-time bound for the knowledge gradient 
policy. Since there are many situations where the outcomes are dichotomous, we 
consider the problem of sequentially making decisions that are rewarded by “successes” 
and “failures”. The binary outcome can be predicted through an unknown 
relationship that depends on partially controllable attributes of each instance. With 
the adaptation of an online Bayesian linear classifier, we design a knowledge gradient 
(KG) policy to guide the experiment. Motivated by personalized medicine where a 
treatment regime is a function that maps individual patient information to a recommended 
treatment, hence explicitly incorporating the heterogeneity in need for 
treatment across individuals, we further extend our knowledge gradient policy to a 
Bayesian contextual bandits setting. Since the sparsity and the relatively small number 
of patients make leaning more dicult, we design an ensemble optimal learning 
method, in which multiple models are strategically generated and combined to minimize 
the incorrect selection of a particularly poorly performing statistical model. 
Driven by numerous needs among materials science society, we developed a KG policy 
for sequential experiments when experiments can be conducted in parallel and/or 
there are multiple tunable parameters which are decided at di↵erent stages in the 
process. Finally, we present a new Modular, Optimal Learning Testing Environment 
(MOLTE) as a public-domain test environment to facilitate the process of more 
comprehensive comparisons, on a broader set of test problems and a broader set of 
policies. 

-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.cs.princeton.edu/pipermail/talks/attachments/20170427/519f027e/attachment.html>


More information about the talks mailing list