releases.shpreview

v0.5.0

GPTQ Quantization, Low-level API

August 22, 2023PEFTView original ↗
$npx -y @buildinternet/releases show rel_-R_OkIeJRDUssK2-8W_PR

GPTQ Integration

Now, you can finetune GPTQ quantized models using PEFT. Here are some examples of how to use PEFT with a GPTQ model: colab notebook and finetuning script.

Low-level API

Enables users and developers to use PEFT as a utility library, at least for injectable adapters (LoRA, IA3, AdaLoRA). It exposes an API to modify the model in place to inject the new layers into the model.

Support for XPU and NPU devices

Leverage the support for more devices for loading and fine-tuning PEFT adapters.

Mix-and-match LoRAs

Stable support and new ways of merging multiple LoRAs. There are currently 3 ways of merging loras supported: linear, svd and cat.

What's Changed

New Contributors

Full Changelog: https://github.com/huggingface/peft/compare/v0.4.0...v0.5.0

Fetched April 7, 2026