This project investigates how sparse weight matrices influence the behavior, efficiency, and performance of neural networks.
We systematically apply different sparsity levels to a fully connected neural network and analyze:
- Training dynamics
- Final model accuracy & generalization
- Computation & memory efficiency
- Gradient flow and stability
The project is part of the Foundations in Deep Learning course and is implemented in PyTorch.
- How does sparsity influence model accuracy?
- Up to which sparsity level can a model perform similarly to a dense network?
- What is the effect on gradient norms, convergence, and stability?
- Does sparsity improve or worsen generalization?
- How much training time and memory can be saved?