Skip to content

Understanding the paper "Principles of Riemannian Geometry in Neural Networks" by Michael Hauser and Asok Ray

Notifications You must be signed in to change notification settings

milosen/riemann-nn

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

9 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Understanding "Principles of Riemannian Geometry in Neural Networks" by Michael Hauser and Asok Ray

This project serves as a playground for experiments inspired by the paper Principles of Riemanian Geometry in Neural Networks.

Main ideas

The experimental setup consists of training a deep neural network with two neurons per layer to separate two classes of data points. Plotting each data point's activations on a 2d scatter plot on every layer helps to gain an understanding of how the data transforms geometrically as it passes through the network. Furthermore, it enables the interpretation of NNs as sequential coordinate transformations, which means that the activation of a single neuron is one coordinate of the respective layer's coordinate system.

An example of this idea can be seen in the following image. The network is a C2-differentiable network as described in the paper, but with the sigmoid activation in the hidden units.

Coordinate transormations of a C2-network

Assuming the output space of a neural network is endowed with a euclidean metric (and assuming some additional technical details), the metric tensor from the last layer can be pulled back through the layers to the input space by

Pullback

with the jakobian

Jacobian

written in tensor notation. This ultimately introduces one possible way of measuring distances between data points in the input space.

In the following, the euclidean metric is pulled back from the last layer to the first and the metric tensors are visualized using their principal components to draw ellipses.

Pullback metric visualized without data points

Pullback metric visualized with data points

About

Understanding the paper "Principles of Riemannian Geometry in Neural Networks" by Michael Hauser and Asok Ray

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published