v1.13.0: Neuron support, IPEX removal, and distributed training fixes
We now have support for AWS Neuron (Trainium/Inferentia) devices. Thanks @michaelbenayoun for adding this.
We've removed IPEX dependency and improved device-agnostic code for XPU.
We've added a bunch of important fixes for FSDP2 users: upcasting only grad-requiring params, better tied embedding errors, DCP optimizer loading, bf16 optimizer step crash fix, and torch < 2.7.0 compatibility.
We've added several fixes to the DeepSpeed + Sequence Parallelism integration introduced in v1.12.0, including evaluation support during SP training and proper process group handling.
We've enhanced FP8 training. Thanks @shimizust for fixing torchao support.
Accelerate now imports faster by deferring heavy dependencies, and torch.compile hooks are disabled lazily.
Fetched April 7, 2026