[Colloquium] REMINDER: 5/22 Talks at TTIC: Yasaman Bahri, Google Brain

Mary Marre via Colloquium colloquium at mailman.cs.uchicago.edu
Mon May 21 16:37:59 CDT 2018


 When:     Tuesday, May 22nd at *11:00 am*

Where:    TTIC, 6045 S Kenwood Avenue, 5th Floor, Room 526

Who:       Yasaman Bahri, Google Brain


Title:        Wide, Deep Neural Networks are Gaussian Processes

Abstract: One means of better understanding seemingly complicated models
such as deep neural networks is to connect them to other objects we already
understand. For instance, Gaussian processes are well-studied models with
well-controlled analytic properties. In his seminal work, Radford Neal
suggested thinking about inference in function space, rather than parameter
space, and in doing so established a correspondence between single-layer
fully-connected neural networks with an i.i.d prior over parameters and
certain Gaussian processes (GPs), in the limit of infinite network width.
The correspondence was, however, restricted to a single-hidden layer.

We develop this line of work and build an exact correspondence between
deep, infinitely wide neural networks and Gaussian processes.
Algorithmically, this mapping also enables a route towards Bayesian
inference with deep neural networks, without needing to instantiate a
network, which we implement on MNIST and CIFAR-10. We compare to the
performance of finite-width networks trained with standard stochastic
optimization. We find that performance increases as finite-width trained
networks are made wider and more similar to a GP, and thus that GP
predictions typically outperform those of finite-width networks.

Time permitting, I will also give some brief highlights of our related
work, studying the propagation of signals through random neural networks.
This analysis informs initializations for training ultra-deep networks with
tens of thousands of layers.

Links:
Jeffrey Pennington and Yasaman Bahri. “Geometry of Neural Network Loss
Surfaces via Random Matrix Theory.” ICML 2017. http://proceedings.mlr.p
ress/v70/pennington17a

Jaehoon Lee*, Yasaman Bahri*, Roman Novak, Samuel S. Schoenholz, Jeffrey
Pennington, Jascha Sohl-Dickstein. “Deep Neural Networks as Gaussian
Processes.” ICLR 2018. https://arxiv.org/abs/1711.00165.

Bio: Yasaman Bahri is a researcher at Google Brain working on deep
learning. The goal of her research is to advance a scientific, principled
understanding of deep learning, with an eye towards theoretical analysis
informed by careful empirical work. She got a PhD in Theoretical Condensed
Matter physics from UC Berkeley, specializing in many body physics; working
on symmetry-protected topological phases, many-body localization, non-Fermi
liquids, and topological mechanics. She is also interested in the
connections between condensed matter, theoretical physics and machine
learning.


Host: Mesrob Ohannessian <mesrob at ttic.edu>




Mary C. Marre
Administrative Assistant
*Toyota Technological Institute*
*6045 S. Kenwood Avenue*
*Room 504*
*Chicago, IL  60637*
*p:(773) 834-1757*
*f: (773) 357-6970*
*mmarre at ttic.edu <mmarre at ttic.edu>*

On Wed, May 16, 2018 at 11:20 AM, Mary Marre <mmarre at ttic.edu> wrote:

> When:     Tuesday, May 22nd at *11:00 am*
>
> Where:    TTIC, 6045 S Kenwood Avenue, 5th Floor, Room 526
>
> Who:       Yasaman Bahri, Google Brain
>
>
> Title:        Wide, Deep Neural Networks are Gaussian Processes
>
> Abstract: One means of better understanding seemingly complicated models
> such as deep neural networks is to connect them to other objects we already
> understand. For instance, Gaussian processes are well-studied models with
> well-controlled analytic properties. In his seminal work, Radford Neal
> suggested thinking about inference in function space, rather than parameter
> space, and in doing so established a correspondence between single-layer
> fully-connected neural networks with an i.i.d prior over parameters and
> certain Gaussian processes (GPs), in the limit of infinite network width.
> The correspondence was, however, restricted to a single-hidden layer.
>
> We develop this line of work and build an exact correspondence between
> deep, infinitely wide neural networks and Gaussian processes.
> Algorithmically, this mapping also enables a route towards Bayesian
> inference with deep neural networks, without needing to instantiate a
> network, which we implement on MNIST and CIFAR-10. We compare to the
> performance of finite-width networks trained with standard stochastic
> optimization. We find that performance increases as finite-width trained
> networks are made wider and more similar to a GP, and thus that GP
> predictions typically outperform those of finite-width networks.
>
> Time permitting, I will also give some brief highlights of our related
> work, studying the propagation of signals through random neural networks.
> This analysis informs initializations for training ultra-deep networks with
> tens of thousands of layers.
>
> Links:
> Jeffrey Pennington and Yasaman Bahri. “Geometry of Neural Network Loss
> Surfaces via Random Matrix Theory.” ICML 2017. http://proceedings.mlr.p
> ress/v70/pennington17a
>
> Jaehoon Lee*, Yasaman Bahri*, Roman Novak, Samuel S. Schoenholz, Jeffrey
> Pennington, Jascha Sohl-Dickstein. “Deep Neural Networks as Gaussian
> Processes.” ICLR 2018. https://arxiv.org/abs/1711.00165.
>
> Bio: Yasaman Bahri is a researcher at Google Brain working on deep
> learning. The goal of her research is to advance a scientific, principled
> understanding of deep learning, with an eye towards theoretical analysis
> informed by careful empirical work. She got a PhD in Theoretical Condensed
> Matter physics from UC Berkeley, specializing in many body physics; working
> on symmetry-protected topological phases, many-body localization, non-Fermi
> liquids, and topological mechanics. She is also interested in the
> connections between condensed matter, theoretical physics and machine
> learning.
>
>
> Host: Mesrob Ohannessian <mesrob at ttic.edu>
>
>
> Mary C. Marre
> Administrative Assistant
> *Toyota Technological Institute*
> *6045 S. Kenwood Avenue*
> *Room 504*
> *Chicago, IL  60637*
> *p:(773) 834-1757*
> *f: (773) 357-6970*
> *mmarre at ttic.edu <mmarre at ttic.edu>*
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://mailman.cs.uchicago.edu/pipermail/colloquium/attachments/20180521/d1d0328b/attachment-0001.html>


More information about the Colloquium mailing list