Skip to content

Release v0.2.7

Choose a tag to compare

@johannaSommer johannaSommer released this 14 Jul 09:46
· 172 commits to main since this release
87eae01

The juiciest bits 🚀

feat: add janus support for quantization+torch.compile combo(s) by @llcnt in #145

You can now decrease the memory impact and the latency of the autoregressive Image Generation model janus(pro-7b) model by quantizing and compiling it.

feat: modular pruning by @begumcig in #154

Pruning is now agnostic to submodules! This recent update enables safe, module-level pruning by finding each target module’s interior, boundary, and exterior, pruning only the interior while auto-patching surrounding shapes.

feat: extend accelerate compatibility by @johannaSommer in #234

We are rolling out more support for base models distributed with accelerate. Several cachers, factorizers and more quantizers are now compatible!

feat: enhance model saving functionality with pro support by @davidberenstein1957 in #200

Saving your smashed models to Huggingface Hub just became easier - we added additional support to better distinguish models that were smashed with pruna and pruna_pro

feat: simplify the setup (tests + dev = dev) by @GreatBahram in #210

@GreatBahram was at it again this release, simplifying the installation and setup for contributors by removing dependency groups - a simple uv sync --extra dev does the job and you can start contributing!

build: reduce core dependencies in pyproject.toml by @ParagEkbote in #227

@ParagEkbote made pruna more lightweight and reduced the package dependencies by 20%!

Pruning some bugs 🐞 and maintenance 🧑‍🌾

New Contributors

Full Changelog: v0.2.6...v0.2.7