This website requires JavaScript.
Explore
Help
Sign In
xinyun
/
vllm
Watch
1
Star
0
Fork
0
You've already forked vllm
mirror of
https://git.datalinker.icu/vllm-project/vllm.git
synced
2026-01-11 09:04:28 +08:00
Code
Issues
Packages
Projects
Releases
Wiki
Activity
vllm
/
vllm
/
engine
/
output_processor
History
zifeitong
974fc9b845
[Bugfix] Fix prompt_logprobs when SamplingParams.detokenize is set to True (
#5226
)
2024-06-04 19:37:28 -07:00
..
__init__.py
[Speculative decoding 6/9] Integrate speculative decoding with LLMEngine (
#3894
)
2024-04-16 13:09:21 -07:00
interfaces.py
[Core] Refactoring sampler and support prompt logprob for chunked prefill (
#4309
)
2024-04-26 13:02:02 +00:00
multi_step.py
[Lora] Support long context lora (
#4787
)
2024-05-18 16:05:23 +09:00
single_step.py
[Bugfix] Fix prompt_logprobs when SamplingParams.detokenize is set to True (
#5226
)
2024-06-04 19:37:28 -07:00
stop_checker.py
[Bugfix] Remove the last EOS token unless explicitly specified (
#5077
)
2024-05-28 17:15:35 -07:00
util.py
[Core] Consolidate prompt arguments to LLM engines (
#4328
)
2024-05-28 13:29:31 -07:00