This website requires JavaScript.
Explore
Help
Sign In
xinyun
/
vllm
Watch
1
Star
0
Fork
0
You've already forked vllm
mirror of
https://git.datalinker.icu/vllm-project/vllm.git
synced
2026-01-25 11:34:32 +08:00
Code
Issues
Packages
Projects
Releases
Wiki
Activity
vllm
/
vllm
/
engine
/
output_processor
History
Cyrus Leung
5ae5ed1e60
[Core] Consolidate prompt arguments to LLM engines (
#4328
)
...
Co-authored-by: Roger Wang <ywang@roblox.com>
2024-05-28 13:29:31 -07:00
..
__init__.py
[Speculative decoding 6/9] Integrate speculative decoding with LLMEngine (
#3894
)
2024-04-16 13:09:21 -07:00
interfaces.py
[Core] Refactoring sampler and support prompt logprob for chunked prefill (
#4309
)
2024-04-26 13:02:02 +00:00
multi_step.py
[Lora] Support long context lora (
#4787
)
2024-05-18 16:05:23 +09:00
single_step.py
[Lora] Support long context lora (
#4787
)
2024-05-18 16:05:23 +09:00
stop_checker.py
[Lora] Support long context lora (
#4787
)
2024-05-18 16:05:23 +09:00
util.py
[Core] Consolidate prompt arguments to LLM engines (
#4328
)
2024-05-28 13:29:31 -07:00