FEAT: Support 1.58-bit LLMs training #114
Labels
enhancement
New feature or request
good first issue
Good for newcomers
help wanted
Extra attention is needed
Low Priority
Hi there!
Microsoft have just released the full handbook for reproduing the 1-bit LLM paper: https://github.com/microsoft/unilm/blob/master/bitnet/The-Era-of-1-bit-LLMs__Training_Tips_Code_FAQ.pdf
Would be exciting to see if we can have an official implementation of that paper in nanotron, and support 1-bit LLM inference directly in transformers for the models that have been trained with that method using nanotron
cc @NouamaneTazi @xrsrke @3outeille @thomwolf
cc original author:
@shumingma
The text was updated successfully, but these errors were encountered: