mirror of
https://git.datalinker.icu/vllm-project/vllm.git
synced 2025-12-15 03:25:01 +08:00
Correct PowerPC to modern IBM Power (#15635)
Signed-off-by: Christy Norman <christy@linux.vnet.ibm.com>
This commit is contained in:
parent
4098b72210
commit
32d669275b
@ -43,7 +43,7 @@ vLLM is flexible and easy to use with:
|
|||||||
- Tensor parallelism and pipeline parallelism support for distributed inference
|
- Tensor parallelism and pipeline parallelism support for distributed inference
|
||||||
- Streaming outputs
|
- Streaming outputs
|
||||||
- OpenAI-compatible API server
|
- OpenAI-compatible API server
|
||||||
- Support NVIDIA GPUs, AMD CPUs and GPUs, Intel CPUs, Gaudi® accelerators and GPUs, PowerPC CPUs, TPU, and AWS Trainium and Inferentia Accelerators.
|
- Support NVIDIA GPUs, AMD CPUs and GPUs, Intel CPUs, Gaudi® accelerators and GPUs, IBM Power CPUs, TPU, and AWS Trainium and Inferentia Accelerators.
|
||||||
- Prefix caching support
|
- Prefix caching support
|
||||||
- Multi-lora support
|
- Multi-lora support
|
||||||
|
|
||||||
|
|||||||
Loading…
x
Reference in New Issue
Block a user