This website requires JavaScript.
Explore
Help
Sign In
xinyun
/
vllm
Watch
1
Star
0
Fork
0
You've already forked vllm
mirror of
https://git.datalinker.icu/vllm-project/vllm.git
synced
2025-12-14 17:15:01 +08:00
Code
Issues
Packages
Projects
Releases
Wiki
Activity
vllm
/
examples
/
offline_inference
History
Iceber Gu
33a3d6c798
fix LoRA-related examples (
#29956
)
...
Signed-off-by: Iceber Gu <caiwei95@hotmail.com>
2025-12-04 11:48:30 +08:00
..
basic
[Chore]:Extract math and argparse utilities to separate modules (
#27188
)
2025-10-26 04:03:32 -07:00
disaggregated-prefill-v1
…
kv_load_failure_recovery
…
logits_processor
…
openai_batch
…
qwen2_5_omni
…
qwen3_omni
…
async_llm_streaming.py
…
audio_language.py
…
automatic_prefix_caching.py
…
batch_llm_inference.py
…
chat_with_tools.py
…
context_extension.py
…
data_parallel.py
…
disaggregated_prefill.py
…
encoder_decoder_multimodal.py
…
llm_engine_example.py
…
llm_engine_reset_kv.py
…
load_sharded_state.py
…
lora_with_quantization_inference.py
fix LoRA-related examples (
#29956
)
2025-12-04 11:48:30 +08:00
metrics.py
…
mistral-small.py
…
mlpspeculator.py
…
multilora_inference.py
fix LoRA-related examples (
#29956
)
2025-12-04 11:48:30 +08:00
prefix_caching.py
…
prompt_embed_inference.py
…
qwen_1m.py
…
reproducibility.py
…
rlhf_colocate.py
…
rlhf_online_quant.py
…
rlhf_utils.py
…
rlhf.py
…
save_sharded_state.py
…
simple_profiling.py
…
skip_loading_weights_in_engine_init.py
…
spec_decode.py
…
structured_outputs.py
…
torchrun_dp_example.py
…
torchrun_example.py
…
vision_language_multi_image.py
…
vision_language.py
…