中文羊驼大模型二期 v1.1 #49
ymcui
announced in
Announcements
Replies: 1 comment
-
Release Note for v1.1This update is primarily focused on functional enhancements, including the addition of scripts supporting FlashAttention-2, vLLM support, and templates for long responses. 🚀 Added FlashAttention-2 training support (#38)
🚀 vLLM acceleration support (#35, #48)
🤖 Template for system prompts for generating long responses (#31)
Other updates
|
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
本次以功能性更新为主,包括添加了FlashAttention-2支持脚本、vLLM支持、长回复模板等。
🚀 已添加FlashAttention-2训练支持(#38)
flash_attn
控制是否开启FlashAttention-2加速训练,具体可查看预训练脚本Wiki或指令精调脚本Wiki🚀 vLLM加速支持(#35、#48)
gradio_demo.py
和inference_hf.py
中添加了对vLLM部署服务的支持。此外,也提供了vLLM作为后端的仿OpenAI API风格的服务器DEMO(openai_api_server_vllm.py
)🤖 生成长回复的系统提示语模板(#31)
alpaca-2-long.txt
系统提示语,可进一步提升回复长度(也可自行尝试修改)其他更新
For English release note, please refer to Discussion.
This discussion was created from the release 中文羊驼大模型二期 v1.1.
Beta Was this translation helpful? Give feedback.
All reactions