Skip to content

v0.0.15: Mistral training, Tensor parallelism improvement, better integration with the AWS SDK

Compare
Choose a tag to compare
@michaelbenayoun michaelbenayoun released this 24 Nov 17:46
· 210 commits to main since this release

What's Changed

Training

Distributed Training

  • parallel_cross_entropy loss support for tensor parallelism (#246)
  • Support for training the Mistral architecture with tensor parallelism (#303)

AWS SDK

  • Fix: neuron_parallel_compile is compatible with the cache system (#352)
  • Full support for neuron_parallel_compile with the cache system: compilation files produced by neuron_parallel_compile will be pushed to the remote cache repo on the Hugging Face Hub at the beginning of the next training job (#354)

Documentation

  • Guide explaining how distributed training works in optimum-neuron (#339)

Inference

  • Data parallelism option for Stable Diffusion - LCM allowing multi-device inference (#346)
  • Support decoding sequences of byte tokens in TGI (#350)

Documentation

  • Updated the documentation on LCM (#351)