Skip to content

Releases: FujitsuResearch/OneCompression

v1.0.2

31 Mar 15:40
bd338e3

Choose a tag to compare

Bug Fix

  • Fixed ImportError when running onecomp CLI without matplotlib installed; AutoBitQuantizer._visualize() now catches the import error and logs a warning instead of crashing

v1.0.1

31 Mar 15:07
637a22c

Choose a tag to compare

Packaging

  • Moved matplotlib from dev extra to new visualize extra in pyproject.toml
  • Made visualize_bit_assignment import lazy in onecomp/quantizer/autobit/__init__.py to avoid requiring matplotlib at import time
  • Updated installation instructions in README.md and docs/getting-started/installation.md to reflect the new visualize extra
  • Updated uv.lock

v1.0.0

31 Mar 14:13
4f982df

Choose a tag to compare

Fujitsu One Compression (OneComp)

A Python package for LLM compression.
Full documentation: https://FujitsuResearch.github.io/OneCompression/

Features

  • Quantization Error Propagation (QEP): Post-training quantization with error propagation to subsequent layers (Arai & Ichikawa, NeurIPS 2025)
  • vLLM Plugin Integration: Serve quantized models with vLLM via built-in DBF and Mixed-GPTQ plugins
  • AutoBit: Mixed-precision quantization with ILP-based bitwidth assignment and automatic VRAM estimation
  • JointQ: Joint quantization optimizing weight assignments and scale parameters simultaneously
  • LoRA SFT Post-Process: Fine-tune quantized models with LoRA for accuracy recovery or knowledge injection
  • Rotation Preprocessing: SpinQuant/OstQuant-based rotation preprocessing (Llama, Qwen3)

Supported Models

Architecture Verified Models Status
Llama TinyLlama, Llama-2, Llama-3 Verified
Qwen3 Qwen3-0.6B ~ 32B Verified

Installation

pip install onecomp