This document summarizes a paper on the emergence of invariance and disentangling in deep representations. The key points are:
1. The paper investigates the relationship between properties desired for representations such as sufficiency, invariance, and disentangling. It shows that under certain model assumptions, minimal sufficiency alone is sufficient to achieve invariance and disentangling.
2. The paper proposes using weight information as a measure of network complexity that can help explain generalization. Weight information is shown to be implicitly minimized by SGD.
3. The paper addresses claims that a new theory of generalization is needed for deep learning by showing weight information recovers the bias-variance tradeoff even when a
Related topics: