v0.18.0: GradientState enhancements and Big Model Inference Fixes
GradientAccumulationPlugin has been added to handle more configurations with the GradientState. Specifically you can optionally disable having Accelerate automatically adjust the length of the scheduler relative to gradient accumulation steps through it. Otherwise Accelerate will now automatically handle ensuring that the schedulers built for non-gradient accumulation will work during gradient accumulationdynamo_backend warning has been silenced.drop_last on linear layers, tied weight loading, and handling of multiple tied parametersfind_tied_parameters now deals with groups of tied parameters (instead of only pairs of them). As a result it now returns a list of list of strings instead of a dictionary.use_orig_params to FullyShardedDataParallelPlugin by @pacman100 in https://github.com/huggingface/accelerate/pull/1184to on modules that wraps accelerate loaded models by @younesbelkada in https://github.com/huggingface/accelerate/pull/1172Full Changelog: https://github.com/huggingface/accelerate/compare/v0.17.1...v0.18.0
Fetched April 7, 2026