Qihan Ren Jiayang Gao Wen Shen Quanshi Zhang†
Shanghai Jiao Tong University
(† Correspondence)
ICLR 2024 [Paper]
This study tries to answer the following question theoretically: can the inference logic of a DNN be explained as symbolic primitives? Although the definition of primitives encoded by a DNN is still an open problem, we follow our recent progress to mathematically formulate primitive patterns using interactions. We have empirically observed the emergence of sparse interaction primitives in many DNNs trained for various tasks, and we note that sparsity is an important characteristic of symbolic representations. In this paper, we further identify three sufficient conditions for the emergence of sparse interaction primitives and prove the sparsity of interactions under such conditions.
Fig.1: Illustration of interaction primitives.
Complexity of an interaction $S$: defined as $|S|$ (also called the order of an interaction)
Interaction primitives: if $|I(S|\boldsymbol{x})|$ is large, we call it a salient interaction primitive; otherwise, if $|I(S|\boldsymbol{x})|$ is close to zero, we call it a noisy pattern.
Understanding interactions as AND relationships: only when patches in $S$ are all present (not masked), the interaction makes a numerical effect $I(S|\boldsymbol{x})$ to the output; otherwise, the numerical effect is removed.
Fig. 2: The sparsity of interaction primitives is observed on different DNNs trained on various datasets.
Sparsity of interaction primitives means that a DNN only encodes a few salient interactions on a specific sample. Only a small number of interactions have significant effects (interactions primitives), while most interactions are noisy patterns.
@inproceedings{
ren2024where,
title={Where We Have Arrived in Proving the Emergence of Sparse Interaction Primitives in {DNN}s},
author={Qihan Ren and Jiayang Gao and Wen Shen and Quanshi Zhang},
booktitle={The Twelfth International Conference on Learning Representations},
year={2024}
}