Search Results for author: Justin Shenk

Found 3 papers, 2 papers with code

Size Matters

no code implementations2 Feb 2021 Mats L. Richter, Wolf Byttner, Ulf Krumnack, Ludwdig Schallner, Justin Shenk

Fully convolutional neural networks can process input of arbitrary size by applying a combination of downsampling and pooling.

Feature Space Saturation during Training

2 code implementations15 Jun 2020 Mats L. Richter, Justin Shenk, Wolf Byttner, Anders Arpteg, Mikael Huss

First, we show that a layer's output can be restricted to the eigenspace of its variance matrix without performance loss.

Spectral Analysis of Latent Representations

1 code implementation19 Jul 2019 Justin Shenk, Mats L. Richter, Anders Arpteg, Mikael Huss

We propose a metric, Layer Saturation, defined as the proportion of the number of eigenvalues needed to explain 99% of the variance of the latent representations, for analyzing the learned representations of neural network layers.

Cannot find the paper you are looking for? You can Submit a new open access paper.