Release v0.2.6
The juiciest bits 🚀
feat: accelerate support by @johannaSommer in #128
Pruna now supports smashing base models that are distributed across several GPUs with accelerate! Enjoy quantizing your big models from two GPUs to just one. We will roll out support for more algorithms as well as compatibility with the EvaluationAgent in the following releases.
feat: switch pruna from poetry to uv by @johnrachwan123 in #164
UV needs no introduction and you can now finally install pruna in lightning speed!
feat: streamline import failure handling by @johannaSommer in #152
We have streamlined the handling of algorithm-specific packages - we verify their correct installation before smashing and now guide the user better through the installation steps if a package is missing.
feat: add dependabot by @GreatBahram in #166
To make sure our dependencies are always up to date and support the newest versions, @GreatBahram introduced a dependabot to the pruna repository! 🤖
feat: improve overall device placement handling by @davidberenstein1957 in #148
To further improve user-experience, we now assist with choosing the best device available for smashing your models.
feat: improve the pre-commit configuration by @GreatBahram in #160
@GreatBahram improved the experience of contributers by overhauling our pre-commit configuration.
feat: update EvaluationAgent to support direct parameters and depreca… by @Ayyanaruto in #188
In their first contribution to the pruna repository, @Ayyanaruto improved the interface of the EvaluationAgent so that users can now directly specify metrics and parameters through the agent's constructor!
Pruning some bugs 🐞 and maintenance 🧑🌾
- fix: torchao rejection test by @johannaSommer in #132
- fix: rewrapping pruna model bug by @begumcig in #174
- fix: pin hqq dependency to avoid model re-loading bug by @davidberenstein1957 in #178
- chore: update pyproject.toml for optional dependencies and bitsandbytes by @davidberenstein1957 in #175
- fix: update poetry setup for external collaborators by @johannaSommer in #191
- docs: improve the documentation by @GreatBahram in #163
- build: update ctranslate2 requirement from ==4.5.0 to ==4.6.0 by @dependabot in #199
- docs: fix note on algorithm argument checking by @johannaSommer in #158
- build: update pytest requirement from 7.4.4 to 8.4.0 by @dependabot in #201
- fix: 168 bug device placement does not work with torchmetric by @davidberenstein1957 in #169
- fix: move import check after availability checks by @johannaSommer in #203
- fix: failing docs tests by @davidberenstein1957 in #139
- docs: change contributors list by @johnrachwan123 in #204
- fix: fix fp8dqrow setting by @nifleisch in #156
- refactor: update function signatures to accept both str and Path types by @Ayyanaruto in #187
- fix: update gptqmodel installation by @johnrachwan123 in #215
- docs: update Flux tutorial as a more general image generation tutorial by @davidberenstein1957 in #127
- docs: update LLM tutorial to optimize and evaluate large language models by @davidberenstein1957 in #126
- fix: code blocks in docs and the code block test by @begumcig in #218
- fix: device of FID torch metric by @johannaSommer in #223
No longer supported 👋
- refactor: remove
torch_staticquantizer 👋 by @johannaSommer in #140
New Contributors
- @dependabot made their first contribution in #193
- @Ayyanaruto made their first contribution in #188
Full Changelog: v0.2.5...v0.2.6