mirror of
https://git.datalinker.icu/vllm-project/vllm.git
synced 2025-12-13 20:44:27 +08:00
Allow send list of str for the Prompt on openai demo endpoint /v1/completions (#323)
* allow str or List[str] for prompt * Update vllm/entrypoints/openai/api_server.py Co-authored-by: Zhuohan Li <zhuohan123@gmail.com> --------- Co-authored-by: Zhuohan Li <zhuohan123@gmail.com>
This commit is contained in:
parent
49b26e2cec
commit
0bd2a573a5
@ -357,6 +357,10 @@ async def create_completion(raw_request: Request):
|
|||||||
|
|
||||||
model_name = request.model
|
model_name = request.model
|
||||||
request_id = f"cmpl-{random_uuid()}"
|
request_id = f"cmpl-{random_uuid()}"
|
||||||
|
if isinstance(request.prompt, list):
|
||||||
|
assert len(request.prompt) == 1
|
||||||
|
prompt = request.prompt[0]
|
||||||
|
else:
|
||||||
prompt = request.prompt
|
prompt = request.prompt
|
||||||
created_time = int(time.time())
|
created_time = int(time.time())
|
||||||
try:
|
try:
|
||||||
|
|||||||
@ -73,7 +73,7 @@ class ChatCompletionRequest(BaseModel):
|
|||||||
|
|
||||||
class CompletionRequest(BaseModel):
|
class CompletionRequest(BaseModel):
|
||||||
model: str
|
model: str
|
||||||
prompt: str
|
prompt: Union[str, List[str]]
|
||||||
suffix: Optional[str] = None
|
suffix: Optional[str] = None
|
||||||
max_tokens: Optional[int] = 16
|
max_tokens: Optional[int] = 16
|
||||||
temperature: Optional[float] = 1.0
|
temperature: Optional[float] = 1.0
|
||||||
|
|||||||
Loading…
x
Reference in New Issue
Block a user