Tag Archives: seminars

Two CBB events: Jeffrey Leek, Computational Biology and Bioinformatics speaker on 5-2-18

I don’t know how many of you have come across this blog, simply statistics https://simplystatistics.org, but he is one of the main contributors, and to say, he’s expert on batch effect and meta-analysis.


CBB_seminar_series_May_ template.pdf

Statseminars Stat & Data Science Seminar, Speaker Carl Zimmer 4/27 @ 11am-1pm

Title: The Library of Babel: On Trying to Read My Genome

Information and Abstract:

Applied Data Science Seminar. Not long ago, information about our DNA was virtually impossible to gain. Now, thanks to the falling cost of DNA sequencing and the growing power of bioinformatics, genetic information is undergoing a Gutenberg-scale explosion of popularity. Millions of people are paying for DNA tests from companies like 23andMe and Ancestry.com, and they are getting unprecedented amounts of information about their ancestry and hereditary diseases. For my latest book, “She Has Her Mother’s Laugh,” I got my genome sequenced and enlisted scientists at Yale and elsewhere to help me interpret it. In my talk, I’ll discuss the results of that exploration–at once enlightening and baffling

For more details and upcoming events visit our website at
http://statistics.yale.edu/ .

YINS Tomorrow – 4/18 Sanjeev Arora, Toward theoretical understanding of deep learning

“Toward theoretical understanding of deep learning”

Speaker: Professor Sanjeev Arora

Princeton University & Institute for Advanced Study

Tomorrow – Wednesday, April 18, 2018, 12:00-1:00pm

Location: Yale Institute for Network Science, 17 Hillhouse Avenue, 3rd floor

Abstract: This talk will be a survey of ongoing efforts and recent results to develop better theoretical understanding of deep learning, from expressiveness to optimization to generalization theory. We will see the (limited) success that has been achieved and the open questions it leads to. (My expository articles appear at
http://www.offconvex.org (link is external))

Bio: Sanjeev Arora is Charles C. Fitzmorris Professor of Computer Science at Princeton University and Visiting Professor at the Institute for Advanced Study. He is an expert in theoretical computer science, especially theoretical ML. He has received the Packard Fellowship (1997), Simons Investigator Award (2012), Goedel Prize (2001 and 2010), ACM-Infosys Foundation Award in the Computing Sciences (now called the ACM prize) (2012), and the Fulkerson Prize in Discrete Math (2012).


4/25/18: Adam Auton (23andme)

5/2/18: Andre Levchenko

Math-applied APPLIED MATH PROGRAM: Seminar & Refreshments Thursday, April 10, 2018


Speaker Mauro Maggioni, John Hopkins University

Date: Tuesday, April 10, 2018

Time: 3:45p.m. Refreshments (AKW, 1st Floor Break Area)

4:00p.m. Seminar (LOM 206)

Title: “Learning and Geometry for Stochastic Dynamical Systems in high dimensions”


We discuss geometry-based statistical learning techniques for performing model reduction and modeling of certain classes of stochastic high-dimensional dynamical systems. We consider two complementary settings. In the first one, we are given long
trajectories of a system, e.g. from molecular dynamics, and we estimate, in a robust fashion, an effective number of degrees of freedom of the system, which may vary in the state space of then system, and a local scale where the dynamics is well-approximated by a reduced dynamics with a small number of degrees of freedom. We then use these ideas to produce an approximation to the generator of the system and obtain, via eigenfunctions of an empirical Fokker-Planck equation (constructed from data), reaction coordinates for the system that capture the large time behavior of the dynamics. We present various examples from molecular dynamics illustrating these ideas.

In the second setting we only have access to a (large number of expensive) simulators that can return short paths of the stochastic system, and introduce a statistical learning framework for estimating local approximations to the system, that can be (automatically) pieced together to form a fast global reduced model for the system, called ATLAS. ATLAS is guaranteed to be accurate (in the sense of producing stochastic paths whose distribution is close to that of paths generated by the original system) not only at small time scales, but also at large time scales, under suitable assumptions on the dynamics. We discuss applications to homogenization of rough diffusions in low and high dimensions, as well as relatively simple systems with separations of time scales, and deterministic chaotic systems in high-dimensions, that are well-approximated by stochastic
diffusion-like equations.
Mauro Maggioni 4-10 flyer.pdf

Seminar by Nobel Laureate W.E. Moerner, April 11th

Attached please find a seminar announcement for Nobel Laureate, W.E. Moerner on Wednesday, April 11, 2018.

Speaker: W.E. Moerner, Nobel Laureate

Title: “Single Molecules for 3D Super-Resolution Imaging and Single Particle Tracking in Cells: Methods and Applications:

Date: Wednesday, April 11, 2018

Time & Place: 3:30 PM, SCL 110

Host: Biophysics Training Grant Students

Admins Please Post

Statseminars Joint Biostatistics / Stat & Data Science Seminar , Speaker Carey E. Priebe, 4/9 @4:15pm-5:30pm


Date: Monday, April 9, 2018

Time: 4:15pm – 5:30pm

Place: Yale Institute for Network Science, 17 Hillhouse Avenue, 3rd Floor, Rm 328

Seminar Speaker: Carey E. Priebe

Department of Applied Mathematics & Statistics, Johns Hopkins University

Personal Website: https://www.ams.jhu.edu/~priebe/

Title: On Spectral Graph Clustering

Abstract: Clustering is a many-splendored thing. As the ill-defined cousin of classification, in which the observation to be classified X comes with a true but unobserved class label Y, clustering is concerned with coherently grouping observations without any explicit concept of true groupings. Spectral graph clustering — clustering the vertices of a graph based on their spectral embedding — is all the rage, and recent theoretical results provide new understanding of the problem and solutions. In particular, we reset the field of spectral graph clustering, demonstrating that spectral graph clustering should not be thought of as kmeans clustering composed with Laplacian spectral embedding, but rather Gaussian mixture model (GMM) clustering composed with either Laplacian or Adjacency spectral embedding (LSE or ASE); in the context of the stochastic blockmodel (SBM), we use eigenvector CLTs & Chernoff analysis to show that (1) GMM dominates kmeans and (2) neither LSE nor ASE dominates, and we present an LSE vs ASE characterization in terms of affinity vs core-periphery SBMs. Along the way, we describe our recent asymptotic efficiency results, as well as an interesting twist on the eigenvector CLT when the block connectivity probability matrix is not positive semidefinite. (And, time permitting, we will touch on essential results using the matrix two-to-infinity norm.) We conclude with a ‘Two Truths’ LSE vs ASE spectral graph clustering result — necessarily including model selection for both embedding dimension & number of clusters — convincingly illustrated via an exciting new diffusion MRI connectome data set: different embedding methods yield different clustering results, with one (ASE) capturing gray matter/white matter separation and the other (LSE) capturing left hemisphere/right hemisphere characterization.

4:00 p.m. Pre-talk Refreshments

4:15 p.m. – 5:30 Seminar, Room 328, 17 Hillhouse Avenue

For more details and upcoming events visit our website at
http://statistics.yale.edu/ .