Revisiting Residual Connections: Orthogonal Updates for Stable and Efficient Deep Networks
Abstract
Orthogonal Residual Updates enhance feature learning and training stability by decomposing module outputs to contribute primarily novel features.
Residual connections are pivotal for deep neural networks, enabling greater depth by mitigating vanishing gradients. However, in standard residual updates, the module's output is directly added to the input stream. This can lead to updates that predominantly reinforce or modulate the existing stream direction, potentially underutilizing the module's capacity for learning entirely novel features. In this work, we introduce Orthogonal Residual Update: we decompose the module's output relative to the input stream and add only the component orthogonal to this stream. This design aims to guide modules to contribute primarily new representational directions, fostering richer feature learning while promoting more efficient training. We demonstrate that our orthogonal update strategy improves generalization accuracy and training stability across diverse architectures (ResNetV2, Vision Transformers) and datasets (CIFARs, TinyImageNet, ImageNet-1k), achieving, for instance, a +4.3\%p top-1 accuracy gain for ViT-B on ImageNet-1k.
Community
๐ Are we using residual connections efficiently in Deep Learning? We propose "Orthogonal Residual Updates": decomposing a module's output relative to the input stream and adding only the component orthogonal to it. This fosters richer feature learning and more efficient training.
๐ป Code: https://github.com/BootsofLagrangian/ortho-residual
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- Deep residual learning with product units (2025)
- VeLU: Variance-enhanced Learning Unit for Deep Neural Networks (2025)
- How to Train Your Metamorphic Deep Neural Network (2025)
- Memory-Efficient Orthogonal Fine-Tuning with Principal Subspace Adaptation (2025)
- CAT: Circular-Convolutional Attention for Sub-Quadratic Transformers (2025)
- Stochastic Variational Propagation: Local, Scalable and Efficient Alternative to Backpropagation (2025)
- Understanding Differential Transformer Unchains Pretrained Self-Attentions (2025)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
Models citing this paper 2
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper