Elena-Alexandra Peste
Graduate School
Alistarh Group
Lampert Group
6 Publications
2023 | Conference Paper | IST-REx-ID: 13053 |
Peste, E.-A., Vladu, A., Kurtic, E., Lampert, C., & Alistarh, D.-A. (n.d.). CrAM: A Compression-Aware Minimizer. In 11th International Conference on Learning Representations . Kigali, Rwanda .
[Preprint]
View
| Files available
| Download Preprint (ext.)
| arXiv
2023 | Thesis | IST-REx-ID: 13074 |
Peste, E.-A. (2023). Efficiency and generalization of sparse neural networks. Institute of Science and Technology Austria. https://doi.org/10.15479/at:ista:13074
[Published Version]
View
| Files available
| DOI
2023 | Conference Paper | IST-REx-ID: 14771 |
Iofinova, E. B., Peste, E.-A., & Alistarh, D.-A. (2023). Bias in pruned vision models: In-depth analysis and countermeasures. In 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition (pp. 24364–24373). Vancouver, BC, Canada: IEEE. https://doi.org/10.1109/cvpr52729.2023.02334
[Preprint]
View
| Files available
| DOI
| Download Preprint (ext.)
| WoS
| arXiv
2022 | Conference Paper | IST-REx-ID: 12299 |
Iofinova, E. B., Peste, E.-A., Kurtz, M., & Alistarh, D.-A. (2022). How well do sparse ImageNet models transfer? In 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (pp. 12256–12266). New Orleans, LA, United States: Institute of Electrical and Electronics Engineers. https://doi.org/10.1109/cvpr52688.2022.01195
[Preprint]
View
| Files available
| DOI
| Download Preprint (ext.)
| WoS
| arXiv
2021 | Journal Article | IST-REx-ID: 10180 |
Hoefler, T., Alistarh, D.-A., Ben-Nun, T., Dryden, N., & Peste, E.-A. (2021). Sparsity in deep learning: Pruning and growth for efficient inference and training in neural networks. Journal of Machine Learning Research. Journal of Machine Learning Research.
[Published Version]
View
| Files available
| Download Published Version (ext.)
| arXiv
2021 | Conference Paper | IST-REx-ID: 11458 |
Peste, E.-A., Iofinova, E. B., Vladu, A., & Alistarh, D.-A. (2021). AC/DC: Alternating Compressed/DeCompressed training of deep neural networks. In 35th Conference on Neural Information Processing Systems (Vol. 34, pp. 8557–8570). Virtual, Online: Curran Associates.
[Published Version]
View
| Files available
| Download Published Version (ext.)
| arXiv
6 Publications
2023 | Conference Paper | IST-REx-ID: 13053 |
Peste, E.-A., Vladu, A., Kurtic, E., Lampert, C., & Alistarh, D.-A. (n.d.). CrAM: A Compression-Aware Minimizer. In 11th International Conference on Learning Representations . Kigali, Rwanda .
[Preprint]
View
| Files available
| Download Preprint (ext.)
| arXiv
2023 | Thesis | IST-REx-ID: 13074 |
Peste, E.-A. (2023). Efficiency and generalization of sparse neural networks. Institute of Science and Technology Austria. https://doi.org/10.15479/at:ista:13074
[Published Version]
View
| Files available
| DOI
2023 | Conference Paper | IST-REx-ID: 14771 |
Iofinova, E. B., Peste, E.-A., & Alistarh, D.-A. (2023). Bias in pruned vision models: In-depth analysis and countermeasures. In 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition (pp. 24364–24373). Vancouver, BC, Canada: IEEE. https://doi.org/10.1109/cvpr52729.2023.02334
[Preprint]
View
| Files available
| DOI
| Download Preprint (ext.)
| WoS
| arXiv
2022 | Conference Paper | IST-REx-ID: 12299 |
Iofinova, E. B., Peste, E.-A., Kurtz, M., & Alistarh, D.-A. (2022). How well do sparse ImageNet models transfer? In 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (pp. 12256–12266). New Orleans, LA, United States: Institute of Electrical and Electronics Engineers. https://doi.org/10.1109/cvpr52688.2022.01195
[Preprint]
View
| Files available
| DOI
| Download Preprint (ext.)
| WoS
| arXiv
2021 | Journal Article | IST-REx-ID: 10180 |
Hoefler, T., Alistarh, D.-A., Ben-Nun, T., Dryden, N., & Peste, E.-A. (2021). Sparsity in deep learning: Pruning and growth for efficient inference and training in neural networks. Journal of Machine Learning Research. Journal of Machine Learning Research.
[Published Version]
View
| Files available
| Download Published Version (ext.)
| arXiv
2021 | Conference Paper | IST-REx-ID: 11458 |
Peste, E.-A., Iofinova, E. B., Vladu, A., & Alistarh, D.-A. (2021). AC/DC: Alternating Compressed/DeCompressed training of deep neural networks. In 35th Conference on Neural Information Processing Systems (Vol. 34, pp. 8557–8570). Virtual, Online: Curran Associates.
[Published Version]
View
| Files available
| Download Published Version (ext.)
| arXiv