[talks] Correction- Colloquium Speaker Erik Sudderth, MONDAY Nov 24, 4:30pm
Nicole E. Wagenblast
nwagenbl at CS.Princeton.EDU
Fri Nov 14 11:19:26 EST 2014
Erik Sudderth, Brown University
Monday, November 24, 2014 - 4:30pm
Computer Science 105
Flexible, Reliable, and Scalable Nonparametric Learning
Applications of statistical machine learning increasingly involve datasets with rich hierarchical, temporal, spatial, or relational structure. Bayesian nonparametric models offer the promise of effective learning from big datasets, but standard inference algorithms often fail in subtle and hard-to-diagnose ways. We explore this issue via variants of a popular and general model family, the hierarchical Dirichlet process. We propose a framework for "memoized" online optimization of variational learning objectives, which achieves computational scalability by processing local batches of data, while simultaneously adapting the global model structure in a coherent fashion. Using this approach, we build improved models of text, audio, image, and social network data.
Erik B. Sudderth is an Assistant Professor in the Brown University Department of Computer Science. He received the Bachelor's degree (summa cum laude, 1999) in Electrical Engineering from the University of California, San Diego, and the Master's and Ph.D. degrees (2006) in EECS from the Massachusetts Institute of Technology. His research interests include probabilistic graphical models; nonparametric Bayesian methods; and applications of statistical machine learning in computer vision and the sciences. He received an NSF CAREER award in 2014, and in 2008 was named one of "AI's 10 to Watch" by IEEE Intelligent Systems Magazine.
talks mailing list
talks at lists.cs.princeton.edu
To edit subscription settings or remove yourself, use this link:
-------------- next part --------------
An HTML attachment was scrubbed...
More information about the talks