Element-wise Modulation of Random Matrices for Efficient Neural Layers
By: Maksymilian Szorc
Fully connected layers are a primary source of memory and computational overhead in deep neural networks due to their dense, often redundant parameterization. While various compression techniques exist, they frequently introduce complex engineering trade-offs or degrade model performance. We propose the Parametrized Random Projection (PRP) layer, a novel approach that decouples feature mixing from adaptation by utilizing a fixed random matrix modulated by lightweight, learnable element-wise parameters. This architecture drastically reduces the trainable parameter count to a linear scale while retaining reliable accuracy across various benchmarks. The design serves as a stable, computationally efficient solution for architectural scaling and deployment in resource-limited settings.
Similar Papers
Beyond Real Weights: Hypercomplex Representations for Stable Quantization
CV and Pattern Recognition
Makes smart AI models smaller and faster.
Pruning Deep Neural Networks via a Combination of the Marchenko-Pastur Distribution and Regularization
Machine Learning (CS)
Makes computer vision models smaller, faster.
Tensor Network Based Feature Learning Model
Machine Learning (CS)
Learns computer patterns faster and better.