ObjectNotFound
|
daf593a385
|
Fix streaming bug (#820)
|
2024-07-30 00:32:07 -07:00 |
|
Yineng Zhang
|
bece265f5a
|
docs: update README (#819)
|
2024-07-30 16:17:50 +10:00 |
|
Liangsheng Yin
|
cdcbde5fc3
|
Code structure refactor (#807)
|
2024-07-29 23:04:48 -07:00 |
|
Enrique Shockwave
|
21e22b9e96
|
Fix LiteLLM kwargs (#817)
|
2024-07-29 22:38:02 -07:00 |
|
Yineng Zhang
|
a50c8a14b3
|
fix: use v0.2.5 for benchmark (#814)
|
2024-07-30 12:40:35 +10:00 |
|
Ying Sheng
|
db6089e6f3
|
Revert "Organize public APIs" (#815)
|
2024-07-29 19:40:28 -07:00 |
|
Liangsheng Yin
|
3520f75fb1
|
Remove inf value for chunked prefill size (#812)
|
2024-07-29 18:34:25 -07:00 |
|
Liangsheng Yin
|
c8e9fed87a
|
Organize public APIs (#809)
|
2024-07-29 15:34:16 -07:00 |
|
yichuan~
|
084fa54d37
|
Add support for OpenAI API : offline batch(file) processing (#699)
Co-authored-by: hnyls2002 <hnyls2002@gmail.com>
|
2024-07-29 13:07:18 -07:00 |
|
Ying Sheng
|
eba458bd19
|
Revert "Revert "fix: update flashinfer to 0.1.2 to fix sampling for cu118"" (#806)
|
2024-07-29 12:20:42 -07:00 |
|
Yineng Zhang
|
3d1cb0af83
|
feat: add chat template for internlm2-chat (#802)
|
2024-07-30 03:18:03 +08:00 |
|
Ying Sheng
|
7d352b4fdd
|
Revert "fix: update flashinfer to 0.1.2 to fix sampling for cu118" (#805)
|
2024-07-29 11:39:12 -07:00 |
|
Yineng Zhang
|
87064015d9
|
fix: update flashinfer to 0.1.2 to fix sampling for cu118 (#803)
|
2024-07-29 11:00:52 -07:00 |
|
Liangsheng Yin
|
7cd4f244a4
|
Chunked prefill (#800)
|
2024-07-29 03:32:58 -07:00 |
|
Ying Sheng
|
98111fbe3e
|
Revert "Chunked prefill support" (#799)
|
2024-07-29 02:38:31 -07:00 |
|
Liangsheng Yin
|
2ec39ab712
|
Chunked prefill support (#797)
|
2024-07-29 02:21:50 -07:00 |
|
ObjectNotFound
|
8f6274c82b
|
Add role documentation, add system begin & end tokens (#793)
|
2024-07-28 23:02:49 -07:00 |
|
Ying Sheng
|
325a06c2de
|
Fix logging (#796)
|
2024-07-28 23:01:45 -07:00 |
|
Ying Sheng
|
79f816292e
|
Fix lazy import location (#795)
|
2024-07-28 22:09:50 -07:00 |
|
Eric Yoon
|
b688fd858d
|
Lazy-import third-party backends (#794)
|
2024-07-28 21:57:41 -07:00 |
|
Ying Sheng
|
5bd899243b
|
Update README.md (#792)
|
2024-07-28 21:57:23 -07:00 |
|
Ying Sheng
|
8d908a937c
|
Fix echo + lobprob for OpenAI API when the prompt is a list (#791)
|
2024-07-28 17:09:16 -07:00 |
|
Yineng Zhang
|
dd7e8b9421
|
chore: add copyright for srt (#790)
|
2024-07-28 23:07:12 +10:00 |
|
Yineng Zhang
|
1f013d64eb
|
docs: make badges center (#789)
|
2024-07-28 22:27:52 +10:00 |
|
Yineng Zhang
|
628e1fa760
|
docs: update README (#788)
|
2024-07-28 22:24:27 +10:00 |
|
Ying Sheng
|
c71880f896
|
Vectorize logprobs computation (#787)
|
2024-07-28 05:22:14 -07:00 |
|
Ying Sheng
|
bcb6611a46
|
Update README.md
|
2024-07-28 01:00:06 -07:00 |
|
Yineng Zhang
|
fa2aa0db0a
|
docs: update index (#786)
|
2024-07-28 17:22:00 +10:00 |
|
Yineng Zhang
|
6a387a69cc
|
fix: exclude logo png in gitignore (#785)
|
2024-07-28 17:08:16 +10:00 |
|
Yineng Zhang
|
27f5ce0a6c
|
fix: init readthedocs support (#784)
|
2024-07-28 16:55:54 +10:00 |
|
Yineng Zhang
|
948625799e
|
docs: init readthedocs support (#783)
|
2024-07-28 16:50:31 +10:00 |
|
Yineng Zhang
|
68e5262699
|
fix: replace pillow with PIL in PACKAGE_LIST (#781)
|
2024-07-28 14:06:24 +10:00 |
|
Lianmin Zheng
|
bc1154c399
|
Bump version to 0.2.6 (#779)
|
2024-07-27 20:29:33 -07:00 |
|
Lianmin Zheng
|
752e643007
|
Allow disabling flashinfer sampling kernel (#778)
|
2024-07-27 20:18:56 -07:00 |
|
Lianmin Zheng
|
30db99b3d9
|
Rename prefill_token_logprobs -> input_token_logprobs; decode_token_logprobs -> output_token_logprobs (#776)
|
2024-07-27 19:50:34 -07:00 |
|
Lianmin Zheng
|
0a409bd438
|
Fix return_log_probs with cuda graph (#775)
|
2024-07-27 19:15:09 -07:00 |
|
Mingyi
|
e4db4e5ba5
|
minor refactor: move check server args to server_args.py (#774)
|
2024-07-27 19:03:40 -07:00 |
|
Lianmin Zheng
|
bbc07c4197
|
Move sampling logits to float32 (#773)
|
2024-07-27 17:30:12 -07:00 |
|
Lianmin Zheng
|
a036d41980
|
Fix max new tokens (#772)
|
2024-07-27 17:22:18 -07:00 |
|
Lianmin Zheng
|
f95e661757
|
Fix max_tokens for OpenAI chat completion API (#766)
|
2024-07-27 15:44:27 -07:00 |
|
Yineng Zhang
|
de854fb5c5
|
feat: add fake tag (#770)
|
2024-07-28 02:22:22 +10:00 |
|
Lianmin Zheng
|
f64b2a9bc0
|
Add slack invitation link.
|
2024-07-27 06:29:15 -07:00 |
|
Ying Sheng
|
9f95dcc64f
|
Update readme (#769)
Co-authored-by: Mingyi <wisclmy0611@gmail.com>
|
2024-07-27 06:12:16 -07:00 |
|
Lianmin Zheng
|
0736b27020
|
[Minor] Improve the code style in TokenizerManager (#767)
|
2024-07-27 05:05:15 -07:00 |
|
Ke Bao
|
3fdab91912
|
Fix TransformerTokenizer init for chatglm2 & 3 (#761)
|
2024-07-27 02:44:46 -07:00 |
|
Liangsheng Yin
|
ba29504b21
|
Update supported models (#763)
|
2024-07-27 15:53:53 +10:00 |
|
Yineng Zhang
|
a72342f180
|
fix: not run workflows on fork repo (#762)
|
2024-07-27 14:51:33 +10:00 |
|
Yineng Zhang
|
c3c74bf874
|
docs: update model support (#760)
|
2024-07-27 14:07:37 +10:00 |
|
Liangsheng Yin
|
d9fccfefe2
|
Fix context length (#757)
|
2024-07-26 18:13:13 -07:00 |
|
Liangsheng Yin
|
679ebcbbdc
|
Deepseek v2 support (#693)
|
2024-07-26 17:10:07 -07:00 |
|