中文羊驼大模型二期 v1.1 #49
ymcui
announced in
Announcements
Replies: 1 comment
-
Release Note for v1.1This update is primarily focused on functional enhancements, including the addition of scripts supporting FlashAttention-2, vLLM support, and templates for long responses. 🚀 Added FlashAttention-2 training support (#38)
🚀 vLLM acceleration support (#35, #48)
🤖 Template for system prompts for generating long responses (#31)
Other updates
|
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
本次以功能性更新为主,包括添加了FlashAttention-2支持脚本、vLLM支持、长回复模板等。
🚀 已添加FlashAttention-2训练支持(#38)
flash_attn控制是否开启FlashAttention-2加速训练,具体可查看预训练脚本Wiki或指令精调脚本Wiki🚀 vLLM加速支持(#35、#48)
gradio_demo.py和inference_hf.py中添加了对vLLM部署服务的支持。此外,也提供了vLLM作为后端的仿OpenAI API风格的服务器DEMO(openai_api_server_vllm.py)🤖 生成长回复的系统提示语模板(#31)
alpaca-2-long.txt系统提示语,可进一步提升回复长度(也可自行尝试修改)其他更新
For English release note, please refer to Discussion.
This discussion was created from the release 中文羊驼大模型二期 v1.1.
Beta Was this translation helpful? Give feedback.
All reactions