This repository was archived by the owner on Mar 17, 2025. It is now read-only.
Commit 94421ea
Merge 1126 (#7)
* Remove hardcode flash-attn disable setting (lm-sys#2342)
* Document turning off proxy_buffering when api is streaming (lm-sys#2337)
* Simplify huggingface api example (lm-sys#2355)
* Update sponsor logos (lm-sys#2367)
* if LOGDIR is empty, then don't try output log to local file (lm-sys#2357)
Signed-off-by: Lei Wen <[email protected]>
Co-authored-by: Lei Wen <[email protected]>
* add best_of and use_beam_search for completions interface (lm-sys#2348)
Signed-off-by: Lei Wen <[email protected]>
Co-authored-by: Lei Wen <[email protected]>
* Extract upvote/downvote from log files (lm-sys#2369)
* Revert "add best_of and use_beam_search for completions interface" (lm-sys#2370)
* Improve doc (lm-sys#2371)
* add best_of and use_beam_search for completions interface (lm-sys#2372)
Signed-off-by: Lei Wen <[email protected]>
Co-authored-by: Lei Wen <[email protected]>
* update monkey patch for llama2 (lm-sys#2379)
* Make E5 adapter more restrict to reduce mismatch (lm-sys#2381)
* Update UI and sponsers (lm-sys#2387)
* Use fsdp api for save save (lm-sys#2390)
* Release v0.2.27
* Spicyboros + airoboros 2.2 template update. (lm-sys#2392)
Co-authored-by: Jon Durbin <[email protected]>
* bugfix of openai_api_server for fastchat.serve.vllm_worker (lm-sys#2398)
Co-authored-by: wuyongyu <[email protected]>
* Revert "bugfix of openai_api_server for fastchat.serve.vllm_worker" (lm-sys#2400)
* Revert "add best_of and use_beam_search for completions interface" (lm-sys#2401)
* Release a v0.2.28 with bug fixes and more test cases
* Fix model_worker error (lm-sys#2404)
* Added google/flan models and fixed AutoModelForSeq2SeqLM when loading T5 compression model (lm-sys#2402)
* Rename twitter to X (lm-sys#2406)
* Update huggingface_api.py (lm-sys#2409)
* Add support for baichuan2 models (lm-sys#2408)
* Fixed character overlap issue when api streaming output (lm-sys#2431)
* Support custom conversation template in multi_model_worker (lm-sys#2434)
* Add Ascend NPU support (lm-sys#2422)
* Add raw conversation template (lm-sys#2417) (lm-sys#2418)
* Improve docs & UI (lm-sys#2436)
* Fix Salesforce xgen inference (lm-sys#2350)
* Add support for Phind-CodeLlama models (lm-sys#2415) (lm-sys#2416)
Co-authored-by: Lianmin Zheng <[email protected]>
* Add falcon 180B chat conversation template (lm-sys#2384)
* Improve docs (lm-sys#2438)
* add dtype and seed (lm-sys#2430)
* Data cleaning scripts for dataset release (lm-sys#2440)
* merge google/flan based adapters: T5Adapter, CodeT5pAdapter, FlanAdapter (lm-sys#2411)
* Fix docs
* Update UI (lm-sys#2446)
* Add Optional SSL Support to controller.py (lm-sys#2448)
* Format & Improve docs
* Release v0.2.29 (lm-sys#2450)
* Show terms of use as an JS alert (lm-sys#2461)
* vllm worker awq quantization update (lm-sys#2463)
Co-authored-by: 董晓龙 <[email protected]>
* Fix falcon chat template (lm-sys#2464)
* Fix chunk handling when partial chunks are returned (lm-sys#2485)
* Update openai_api_server.py to add an SSL option (lm-sys#2484)
* Update vllm_worker.py (lm-sys#2482)
* fix typo quantization (lm-sys#2469)
* fix vllm quanziation args
* Update README.md (lm-sys#2492)
* Huggingface api worker (lm-sys#2456)
* Update links to lmsys-chat-1m (lm-sys#2497)
* Update train code to support the new tokenizer (lm-sys#2498)
* Third Party UI Example (lm-sys#2499)
* Add metharme (pygmalion) conversation template (lm-sys#2500)
* Optimize for proper flash attn causal handling (lm-sys#2503)
* Add Mistral AI instruction template (lm-sys#2483)
* Update monitor & plots (lm-sys#2506)
* Release v0.2.30 (lm-sys#2507)
* Fix for single turn dataset (lm-sys#2509)
* replace os.getenv with os.path.expanduser because the first one doesn… (lm-sys#2515)
Co-authored-by: khalil <[email protected]>
* Fix arena (lm-sys#2522)
* Update Dockerfile (lm-sys#2524)
* add Llama2ChangAdapter (lm-sys#2510)
* Add ExllamaV2 Inference Framework Support. (lm-sys#2455)
* Improve docs (lm-sys#2534)
* Fix warnings for new gradio versions (lm-sys#2538)
* revert the gradio change; now works for 3.40
* Improve chat templates (lm-sys#2539)
* Add Zephyr 7B Alpha (lm-sys#2535)
* Improve Support for Mistral-Instruct (lm-sys#2547)
* correct max_tokens by context_length instead of raise exception (lm-sys#2544)
* Revert "Improve Support for Mistral-Instruct" (lm-sys#2552)
* Fix Mistral template (lm-sys#2529)
* Add additional Informations from the vllm worker (lm-sys#2550)
* Make FastChat work with LMSYS-Chat-1M Code (lm-sys#2551)
* Create `tags` attribute to fix `MarkupError` in rich CLI (lm-sys#2553)
* move BaseModelWorker outside serve.model_worker to make it independent (lm-sys#2531)
* Misc style and bug fixes (lm-sys#2559)
* Fix README.md (lm-sys#2561)
* release v0.2.31 (lm-sys#2563)
* resolves lm-sys#2542 modify dockerfile to upgrade cuda to 12.2.0 and pydantic 1.10.13 (lm-sys#2565)
* Add airoboros_v3 chat template (llama-2 format) (lm-sys#2564)
* Add Xwin-LM V0.1, V0.2 support (lm-sys#2566)
* Fixed model_worker generate_gate may blocked main thread (lm-sys#2540) (lm-sys#2562)
* feat: add claude-v2 (lm-sys#2571)
* Update vigogne template (lm-sys#2580)
* Fix issue lm-sys#2568: --device mps led to TypeError: forward() got an unexpected keyword argument 'padding_mask'. (lm-sys#2579)
* Add Mistral-7B-OpenOrca conversation_temmplate (lm-sys#2585)
* docs: bit misspell comments model adapter default template name conversation (lm-sys#2594)
* Update Mistral template (lm-sys#2581)
* Fix <s> in mistral template
* Update README.md (vicuna-v1.3 -> vicuna-1.5) (lm-sys#2592)
* Update README.md to highlight chatbot arena (lm-sys#2596)
* Add Lemur model (lm-sys#2584)
Co-authored-by: Roberto Ugolotti <[email protected]>
* add trust_remote_code=True in BaseModelAdapter (lm-sys#2583)
* Openai interface add use beam search and best of 2 (lm-sys#2442)
Signed-off-by: Lei Wen <[email protected]>
Co-authored-by: Lei Wen <[email protected]>
* Update qwen and add pygmalion (lm-sys#2607)
* feat: Support model AquilaChat2 (lm-sys#2616)
* Added settings vllm (lm-sys#2599)
Co-authored-by: bodza <[email protected]>
Co-authored-by: bodza <[email protected]>
* [Logprobs] Support logprobs=1 (lm-sys#2612)
* release v0.2.32
* fix: Fix for OpenOrcaAdapter to return correct conversation template (lm-sys#2613)
* Make fastchat.serve.model_worker to take debug argument (lm-sys#2628)
Co-authored-by: hi-jin <[email protected]>
* openchat 3.5 model support (lm-sys#2638)
* xFastTransformer framework support (lm-sys#2615)
* feat: support custom models vllm serving (lm-sys#2635)
* kill only fastchat process (lm-sys#2641)
* Update server_arch.png
* Use conv.update_last_message api in mt-bench answer generation (lm-sys#2647)
* Improve Azure OpenAI interface (lm-sys#2651)
* Add required_temp support in jsonl format to support flexible temperature setting for gen_api_answer (lm-sys#2653)
* Pin openai version < 1 (lm-sys#2658)
* Remove exclude_unset parameter (lm-sys#2654)
* Revert "Remove exclude_unset parameter" (lm-sys#2666)
* added support for CodeGeex(2) (lm-sys#2645)
* add chatglm3 conv template support in conversation.py (lm-sys#2622)
* UI and model change (lm-sys#2672)
Co-authored-by: Lianmin Zheng <[email protected]>
* train_flant5: fix typo (lm-sys#2673)
* Fix gpt template (lm-sys#2674)
* Update README.md (lm-sys#2679)
* feat: support template's stop_str as list (lm-sys#2678)
* Update exllama_v2.md (lm-sys#2680)
* save model under deepspeed (lm-sys#2689)
* Adding SSL support for model workers and huggingface worker (lm-sys#2687)
* Check the max_new_tokens <= 0 in openai api server (lm-sys#2688)
* Add Microsoft/Orca-2-7b and update model support docs (lm-sys#2714)
* fix tokenizer of chatglm2 (lm-sys#2711)
* Template for using Deepseek code models (lm-sys#2705)
* add support for Chinese-LLaMA-Alpaca (lm-sys#2700)
* Make --load-8bit flag work with weights in safetensors format (lm-sys#2698)
* Format code and minor bug fix (lm-sys#2716)
* Bump version to v0.2.33 (lm-sys#2717)
* fix tokenizer.pad_token attribute error (lm-sys#2710)
* support stable-vicuna model (lm-sys#2696)
* Exllama cache 8bit (lm-sys#2719)
* Add Yi support (lm-sys#2723)
* Add Hermes 2.5 [fixed] (lm-sys#2725)
* Fix Hermes2Adapter (lm-sys#2727)
* Fix YiAdapter (lm-sys#2730)
* add trust_remote_code argument (lm-sys#2715)
* Add revision arg to MT Bench answer generation (lm-sys#2728)
* Fix MPS backend 'index out of range' error (lm-sys#2737)
* add starling support (lm-sys#2738)
---------
Signed-off-by: Lei Wen <[email protected]>
Co-authored-by: Trangle <[email protected]>
Co-authored-by: Nathan Stitt <[email protected]>
Co-authored-by: Lianmin Zheng <[email protected]>
Co-authored-by: leiwen83 <[email protected]>
Co-authored-by: Lei Wen <[email protected]>
Co-authored-by: Jon Durbin <[email protected]>
Co-authored-by: Jon Durbin <[email protected]>
Co-authored-by: Rayrtfr <[email protected]>
Co-authored-by: wuyongyu <[email protected]>
Co-authored-by: wangxiyuan <[email protected]>
Co-authored-by: Jeff (Zhen) Wang <[email protected]>
Co-authored-by: karshPrime <[email protected]>
Co-authored-by: obitolyz <[email protected]>
Co-authored-by: Shangwei Chen <[email protected]>
Co-authored-by: HyungJin Ahn <[email protected]>
Co-authored-by: zhangsibo1129 <[email protected]>
Co-authored-by: Tobias Birchler <[email protected]>
Co-authored-by: Jae-Won Chung <[email protected]>
Co-authored-by: Mingdao Liu <[email protected]>
Co-authored-by: Ying Sheng <[email protected]>
Co-authored-by: Brandon Biggs <[email protected]>
Co-authored-by: dongxiaolong <[email protected]>
Co-authored-by: 董晓龙 <[email protected]>
Co-authored-by: Siddartha Naidu <[email protected]>
Co-authored-by: shuishu <[email protected]>
Co-authored-by: Andrew Aikawa <[email protected]>
Co-authored-by: Liangsheng Yin <[email protected]>
Co-authored-by: enochlev <[email protected]>
Co-authored-by: AlpinDale <[email protected]>
Co-authored-by: Lé <[email protected]>
Co-authored-by: Toshiki Kataoka <[email protected]>
Co-authored-by: khalil <[email protected]>
Co-authored-by: khalil <[email protected]>
Co-authored-by: dubaoquan404 <[email protected]>
Co-authored-by: Chang W. Lee <[email protected]>
Co-authored-by: theScotchGame <[email protected]>
Co-authored-by: lewtun <[email protected]>
Co-authored-by: Stephen Horvath <[email protected]>
Co-authored-by: liunux4odoo <[email protected]>
Co-authored-by: Norman Mu <[email protected]>
Co-authored-by: Sebastian Bodza <[email protected]>
Co-authored-by: Tianle (Tim) Li <[email protected]>
Co-authored-by: Wei-Lin Chiang <[email protected]>
Co-authored-by: Alex <[email protected]>
Co-authored-by: Jingcheng Hu <[email protected]>
Co-authored-by: lvxuan <[email protected]>
Co-authored-by: cOng <[email protected]>
Co-authored-by: bofeng huang <[email protected]>
Co-authored-by: Phil-U-U <[email protected]>
Co-authored-by: Wayne Spangenberg <[email protected]>
Co-authored-by: Guspan Tanadi <[email protected]>
Co-authored-by: Rohan Gupta <[email protected]>
Co-authored-by: ugolotti <[email protected]>
Co-authored-by: Roberto Ugolotti <[email protected]>
Co-authored-by: edisonwd <[email protected]>
Co-authored-by: FangYin Cheng <[email protected]>
Co-authored-by: bodza <[email protected]>
Co-authored-by: bodza <[email protected]>
Co-authored-by: Cody Yu <[email protected]>
Co-authored-by: Srinath Janakiraman <[email protected]>
Co-authored-by: Jaeheon Jeong <[email protected]>
Co-authored-by: One <[email protected]>
Co-authored-by: [email protected] <[email protected]>
Co-authored-by: David <[email protected]>
Co-authored-by: Witold Wasiczko <[email protected]>
Co-authored-by: Peter Willemsen <[email protected]>
Co-authored-by: ZeyuTeng96 <[email protected]>
Co-authored-by: Forceless <[email protected]>
Co-authored-by: Jeff <[email protected]>
Co-authored-by: MrZhengXin <[email protected]>
Co-authored-by: Long Nguyen <[email protected]>
Co-authored-by: Elsa Granger <[email protected]>
Co-authored-by: Christopher Chou <[email protected]>
Co-authored-by: wangshuai09 <[email protected]>
Co-authored-by: amaleshvemula <[email protected]>
Co-authored-by: Zollty Tsou <[email protected]>
Co-authored-by: xuguodong1999 <[email protected]>
Co-authored-by: Michael J Kaye <[email protected]>
Co-authored-by: 152334H <[email protected]>
Co-authored-by: Jingsong-Yan <[email protected]>
Co-authored-by: Siyuan (Ryans) Zhuang <[email protected]>1 parent a887de7 commit 94421ea
File tree
62 files changed
+6801
-5987
lines changed- assets
- data
- docker
- docs
- commands
- fastchat
- llm_judge
- model
- modules
- protocol
- serve
- monitor
- dataset_release_scripts/lmsys_chat_1m
- train
- scripts
- tests
Some content is hidden
Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.
62 files changed
+6801
-5987
lines changedLoading
Large diffs are not rendered by default.
| Original file line number | Diff line number | Diff line change | |
|---|---|---|---|
| |||
1 | | - | |
| 1 | + | |
2 | 2 | | |
3 | 3 | | |
4 | 4 | | |
5 | 5 | | |
6 | | - | |
| 6 | + | |
| 7 | + | |
| Original file line number | Diff line number | Diff line change | |
|---|---|---|---|
| |||
23 | 23 | | |
24 | 24 | | |
25 | 25 | | |
26 | | - | |
| 26 | + | |
27 | 27 | | |
28 | 28 | | |
29 | 29 | | |
| |||
| Original file line number | Diff line number | Diff line change | |
|---|---|---|---|
| |||
24 | 24 | | |
25 | 25 | | |
26 | 26 | | |
| 27 | + | |
| 28 | + | |
| 29 | + | |
| 30 | + | |
| 31 | + | |
| 32 | + | |
| 33 | + | |
| 34 | + | |
| 35 | + | |
| 36 | + | |
| 37 | + | |
| Original file line number | Diff line number | Diff line change | |
|---|---|---|---|
| |||
72 | 72 | | |
73 | 73 | | |
74 | 74 | | |
75 | | - | |
| 75 | + | |
| 76 | + | |
| 77 | + | |
| 78 | + | |
| 79 | + | |
| 80 | + | |
| 81 | + | |
| 82 | + | |
| 83 | + | |
| 84 | + | |
76 | 85 | | |
77 | 86 | | |
78 | 87 | | |
| |||
| Original file line number | Diff line number | Diff line change | |
|---|---|---|---|
| |||
| 1 | + | |
| 2 | + | |
| 3 | + | |
| 4 | + | |
| 5 | + | |
| 6 | + | |
| Original file line number | Diff line number | Diff line change | |
|---|---|---|---|
| |||
| 1 | + | |
| 2 | + | |
| 3 | + | |
| 4 | + | |
| 5 | + | |
| 6 | + | |
| 7 | + | |
| 8 | + | |
| 9 | + | |
| 10 | + | |
| 11 | + | |
| 12 | + | |
| 13 | + | |
| 14 | + | |
| 15 | + | |
| 16 | + | |
| 17 | + | |
| 18 | + | |
| 19 | + | |
| 20 | + | |
| 21 | + | |
| 22 | + | |
| 23 | + | |
| 24 | + | |
| 25 | + | |
| 26 | + | |
| 27 | + | |
| 28 | + | |
| 29 | + | |
| 30 | + | |
| 31 | + | |
| 32 | + | |
| 33 | + | |
| 34 | + | |
| 35 | + | |
| 36 | + | |
| 37 | + | |
| 38 | + | |
| 39 | + | |
| 40 | + | |
| 41 | + | |
| 42 | + | |
| 43 | + | |
| 44 | + | |
| 45 | + | |
| 46 | + | |
| 47 | + | |
| 48 | + | |
| 49 | + | |
| 50 | + | |
| 51 | + | |
| 52 | + | |
| 53 | + | |
| 54 | + | |
| 55 | + | |
| 56 | + | |
| 57 | + | |
| 58 | + | |
| 59 | + | |
| 60 | + | |
| 61 | + | |
| 62 | + | |
| 63 | + | |
| Original file line number | Diff line number | Diff line change | |
|---|---|---|---|
| |||
19 | 19 | | |
20 | 20 | | |
21 | 21 | | |
22 | | - | |
| 22 | + | |
23 | 23 | | |
24 | 24 | | |
25 | 25 | | |
| |||
| Original file line number | Diff line number | Diff line change | |
|---|---|---|---|
| |||
5 | 5 | | |
6 | 6 | | |
7 | 7 | | |
8 | | - | |
| 8 | + | |
9 | 9 | | |
| 10 | + | |
| 11 | + | |
10 | 12 | | |
11 | 13 | | |
12 | 14 | | |
| |||
30 | 32 | | |
31 | 33 | | |
32 | 34 | | |
| 35 | + | |
| 36 | + | |
33 | 37 | | |
34 | 38 | | |
35 | 39 | | |
| |||
45 | 49 | | |
46 | 50 | | |
47 | 51 | | |
| 52 | + | |
| 53 | + | |
| 54 | + | |
| 55 | + | |
| 56 | + | |
48 | 57 | | |
49 | 58 | | |
50 | 59 | | |
| |||
64 | 73 | | |
65 | 74 | | |
66 | 75 | | |
67 | | - | |
| 76 | + | |
68 | 77 | | |
69 | 78 | | |
70 | 79 | | |
| |||
0 commit comments