Skip to yearly menu bar Skip to main content


Poster

Weight-Sharing Regularization

Mehran Shakerinava · Motahareh Sohrabi · Siamak Ravanbakhsh · Simon Lacoste-Julien

MR1 & MR2 - Number 136

Abstract: Weight-sharing is ubiquitous in deep learning. Motivated by this, we propose a weight-sharing regularization'' penalty on the weights wRdwRd of a neural network, defined as R(w)=1d1di>j|wiwj|R(w)=1d1di>j|wiwj|. We study the proximal mapping of RR and provide an intuitive interpretation of it in terms of a physical system of interacting particles. We also parallelize existing algorithms for proxRproxR (to run on GPU) and find that one of them is fast in practice but slow (O(d)O(d)) for worst-case inputs. Using the physical interpretation, we design a novel parallel algorithm which runs in O(log3d)O(log3d) when sufficient processors are available, thus guaranteeing fast training. Our experiments reveal that weight-sharing regularization enables fully connected networks to learn convolution-like filters even when pixels have been shuffled while convolutional neural networks fail in this setting. Our code is available on \href{https://github.com/motahareh-sohrabi/weight-sharing-regularization}{github}.

Chat is not available.