Skip to content

v0.20.0: MPS and fp4 support on Big Model Inference, 4-bit QLoRA, Intel GPU, Distributed Inference, and much more!

Compare
Choose a tag to compare
@sgugger sgugger released this 07 Jun 19:33
· 771 commits to main since this release
9765b84

Big model inference

Support has been added to run device_map="auto" on the MPS device. Big model inference also work with models loaded in 4 bits in Transformers.

4-bit QLoRA Support

Distributed Inference Utilities

This version introduces a new Accelerator.split_between_processes utility to help with performing distributed infernece with non-tensorized or non-dataloader workflows. Read more here

Introduce XPU support for Intel GPU

Add support for the new PyTorch XLA TPU runtime

  • Accelerate now supports the latest TPU runtimes #1393, #1385

A new optimizer method: LocalSGD

  • This is a new wrapper around SGD which enables efficient multi-GPU training in the case when no fast interconnect is possible by @searchivarius in #1378

Papers with 🤗 Accelerate

  • We now have an entire section of the docs dedicated to official paper implementations and citations using the framework #1399, see it live here

Breaking changes

logging_dir has been fully deprecated, please use project_dir or a Project_configuration

What's new?

Significant community contributions

The following contributors have made significant changes to the library over the last release:

  • @will-cromar
    • Support TPU v4 with new PyTorch/XLA TPU runtime (#1393)
    • Support TPU v2 and v3 on new PyTorch/XLA TPU runtime (#1385)
  • @searchivarius
    • Adding support for local SGD. (#1378)
  • @abhilash1910
    • Intel GPU support initialization (#1118)
    • Fix bug on ipex for diffusers (#1426)
    • Refactor and simplify xpu device in state (#1456)
    • NVME path support for deepspeed (#1484)
  • @sywangyi
    • fix error for CPU DDP using trainer api. (#1455)
    • fix crash when ipex is installed and torch has no xpu (#1502)
    • should set correct dtype to ipex optimize and use amp logic in native… (#1511)
    • remove ipexplugin, let ACCELERATE_USE_IPEX/ACCELERATE_USE_XPU control the ipex and xpu (#1503)
  • @Ethan-yt
    • Fix gradient state bugs in multiple dataloader (#1483)