releases.shpreview

v0.2.0

March 10, 2023PEFTView original ↗
$npx -y @buildinternet/releases show rel_ABpfLY3R85fHJuITkUYxS

Whisper large tuning using PEFT LoRA+INT-8 on T4 GPU in Colab notebooks

We tested PEFT on @OpenAI's Whisper Large model and got: i) 5x larger batch sizes ii) Less than 8GB GPU VRAM iii) Best part? Almost no degredation to WER 🤯

Without PEFT:

  • OOM on a T4 GPU ❌
  • 6GB checkpoint ❌
  • 13.64 WER ✅

With PEFT:

  • Train on a T4 GPU ✅
  • 60MB checkpoint ✅
  • 14.01 WER ✅

prepare_for_int8_training utility

This utility enables preprocessing the base model to be ready for INT8 training.

disable_adapter() context manager

Enables to disable adapter layers to get the outputs from the frozen base models. An exciting application of this feature allows only a single model copy to be used for policy model and reference model generations in RLHF.

What's Changed

New Contributors

Significant community contributions

The following contributors have made significant changes to the library over the last release:

  • @mayank31398
    • Prompt Tuning method enhancements and fixes (#68, #150)

Full Changelog: https://github.com/huggingface/peft/compare/v0.1.0...v0.2.0

Fetched April 7, 2026