This website requires JavaScript.
Explore
Help
Sign In
xinyun
/
vllm
Watch
1
Star
0
Fork
0
You've already forked vllm
mirror of
https://git.datalinker.icu/vllm-project/vllm.git
synced
2026-01-11 20:24:34 +08:00
Code
Issues
Packages
Projects
Releases
Wiki
Activity
vllm
/
vllm
/
engine
History
Cyrus Leung
7eb4a51c5f
[Core] Support serving encoder/decoder models (
#7258
)
2024-08-09 10:39:41 +08:00
..
output_processor
[core][misc] simply output processing with shortcut code path (
#7117
)
2024-08-04 00:22:19 -07:00
__init__.py
Change the name to vLLM (
#150
)
2023-06-17 03:07:40 -07:00
arg_utils.py
[Core] Subclass ModelRunner to support cross-attention & encoder sequences (towards eventual encoder/decoder model support) (
#4942
)
2024-08-06 16:51:47 -04:00
async_llm_engine.py
[Core] Support serving encoder/decoder models (
#7258
)
2024-08-09 10:39:41 +08:00
async_timeout.py
[Bugfix] AsyncLLMEngine hangs with asyncio.run (
#5654
)
2024-06-19 13:57:12 -07:00
llm_engine.py
[Core] Support serving encoder/decoder models (
#7258
)
2024-08-09 10:39:41 +08:00
metrics.py
[Bugfix] StatLoggers: cache spec decode metrics when they get collected. (
#6645
)
2024-07-23 23:05:05 +00:00
protocol.py
[BugFix] Overhaul async request cancellation (
#7111
)
2024-08-07 13:21:41 +08:00