v0.4.0: peft integrationApply RLHF and fine-tune your favorite large model on consumer GPU using peft and trl ! Share also easily your trained RLHF adapters on the Hub with few lines of code
With this integration you can train gpt-neo-x (20B parameter model - 40GB in bfloat16) on a 24GB consumer GPU!
core] Fix quality issue by @younesbelkada in https://github.com/lvwerra/trl/pull/197peft integration by @edbeeching in https://github.com/lvwerra/trl/pull/163core] Update dependency by @younesbelkada in https://github.com/lvwerra/trl/pull/206Full Changelog: https://github.com/lvwerra/trl/compare/v0.3.1...v0.4.0
Fetched April 7, 2026