
A dataset full of simple yet diverse stories; the MNIST for language

This paper presents a comprehensive taxonomy of interpretation and explanation methods developed for capsule network (capsnet) architectures, analyzing their mechanisms, applicability, and performance across diverse problem domains.

Decomposing activations into sparse polynomials and using their geometry

Sim makes post-hoc interpretability tools more effective through latent space constraints

We show the applicability of twin network augmentation on convolutional neural networks

We introduce a label-efficient approach for radio frequency fingerprint identification, achieving competitive accuracy with up to 10x fewer labels.

This paper provides a systematic and principled study on the interpretability of capsule network (capsnet) representations, aiming to characterize the nature and structure of the learned features across diverse architectures and datasets

This study investigates the internal mechanisms of matrix capsule networks with the EM routing algorithm

Introducing a global svd-like algorithm for multi-linear models.

Using bilinear MLPs to reverse-engineer shallow MNIST and tiny stories models from their weights.

A novel training strategy for improved spiking neural networks and efficient weight quantization.

We use a per-token bias in SAEs to separate token reconstructions from interesting features.

Introducing bilinear MLPs as a new approach to weight-based interpretability.

Self-supervised contrastive learning for multi-label hyperspectral image classification

A deep learning model for hyperspectral remote sensing, shifting from traditional single-label, pixel-level classification to multi-label, patch-level analysis

Three techniques to significantly improve the forward-forward algorithm. We achieve 84% on CIFAR-10.

This study presents a deep learning model for arabic aspect-based sentiment analysis (ABSA) using gated recurrent units (gru) combined with features from the multilingual universal sentence encoder (muse)