Logo
Explore Help
Sign In
xinyun/vllm
1
0
Fork 0
You've already forked vllm
mirror of https://git.datalinker.icu/vllm-project/vllm.git synced 2026-01-11 09:04:28 +08:00
Code Issues Packages Projects Releases Wiki Activity
vllm/vllm/engine/output_processor
History
zifeitong 974fc9b845
[Bugfix] Fix prompt_logprobs when SamplingParams.detokenize is set to True (#5226)
2024-06-04 19:37:28 -07:00
..
__init__.py
[Speculative decoding 6/9] Integrate speculative decoding with LLMEngine (#3894)
2024-04-16 13:09:21 -07:00
interfaces.py
[Core] Refactoring sampler and support prompt logprob for chunked prefill (#4309)
2024-04-26 13:02:02 +00:00
multi_step.py
[Lora] Support long context lora (#4787)
2024-05-18 16:05:23 +09:00
single_step.py
[Bugfix] Fix prompt_logprobs when SamplingParams.detokenize is set to True (#5226)
2024-06-04 19:37:28 -07:00
stop_checker.py
[Bugfix] Remove the last EOS token unless explicitly specified (#5077)
2024-05-28 17:15:35 -07:00
util.py
[Core] Consolidate prompt arguments to LLM engines (#4328)
2024-05-28 13:29:31 -07:00
Powered by Gitea Version: 1.23.1 Page: 537ms Template: 3ms
English
Bahasa Indonesia Deutsch English Español Français Gaeilge Italiano Latviešu Magyar nyelv Nederlands Polski Português de Portugal Português do Brasil Suomi Svenska Türkçe Čeština Ελληνικά Български Русский Українська فارسی മലയാളം 日本語 简体中文 繁體中文(台灣) 繁體中文(香港) 한국어
Licenses API