https://openreview.net/forum?id=OWmu3QOa0O&referrer=%5Bthe%20profile%20of%20Nolan%20Simran%20Dey%5D(%2Fprofile%3Fid%3D~Nolan_Simran_Dey1)
Several challenges make it difficult for sparse neural networks to compete with dense models. First, setting a large fraction of weights to zero impairs...
holistic approachsparsemaximalupdateparameterization
https://openreview.net/forum?id=HddmvY8XvEt&referrer=%5Bthe%20profile%20of%20Mike%20Lasby%5D(%2Fprofile%3Fid%3D~Mike_Lasby1)
Dynamic Sparse Training (DST) methods achieve state-of-the-art results in sparse neural network training, matching the generalization of dense models while...
sparse trainingdynamicstructuredsparsityopenreview
https://openreview.net/forum?id=WDgV1BJEW0&referrer=%5Bthe%20profile%20of%20Yongduo%20Sui%5D(%2Fprofile%3Fid%3D~Yongduo_Sui1)
Graph Neural Networks (GNNs) excel in various graph learning tasks but face computational challenges when applied to large-scale graphs. A promising solution...
two headssparse trainingbetteroneboosting
https://openreview.net/forum?id=8abNCVJs2j
Training deep neural networks (DNNs) is costly. Fortunately, Nvidia Ampere and Hopper GPUs can accelerate matrix multiplications twice as fast as a dense...
stecontinuouspruningfunctionefficient