mirror of
https://git.datalinker.icu/vllm-project/vllm.git
synced 2025-12-13 15:35:01 +08:00
chore(examples-docs): upgrade to OpenAI V1 (#1785)
This commit is contained in:
parent
cd3aa153a4
commit
c07a442854
@ -157,11 +157,16 @@ Since this server is compatible with OpenAI API, you can use it as a drop-in rep
|
|||||||
|
|
||||||
.. code-block:: python
|
.. code-block:: python
|
||||||
|
|
||||||
import openai
|
from openai import OpenAI
|
||||||
|
|
||||||
# Modify OpenAI's API key and API base to use vLLM's API server.
|
# Modify OpenAI's API key and API base to use vLLM's API server.
|
||||||
openai.api_key = "EMPTY"
|
openai_api_key = "EMPTY"
|
||||||
openai.api_base = "http://localhost:8000/v1"
|
openai_api_base = "http://localhost:8000/v1"
|
||||||
completion = openai.Completion.create(model="facebook/opt-125m",
|
client = OpenAI(
|
||||||
|
api_key=openai_api_key,
|
||||||
|
base_url=openai_api_base,
|
||||||
|
)
|
||||||
|
completion = client.completions.create(model="facebook/opt-125m",
|
||||||
prompt="San Francisco is a")
|
prompt="San Francisco is a")
|
||||||
print("Completion result:", completion)
|
print("Completion result:", completion)
|
||||||
|
|
||||||
@ -194,11 +199,17 @@ Using the `openai` python package, you can also communicate with the model in a
|
|||||||
|
|
||||||
.. code-block:: python
|
.. code-block:: python
|
||||||
|
|
||||||
import openai
|
from openai import OpenAI
|
||||||
# Set OpenAI's API key and API base to use vLLM's API server.
|
# Set OpenAI's API key and API base to use vLLM's API server.
|
||||||
openai.api_key = "EMPTY"
|
openai_api_key = "EMPTY"
|
||||||
openai.api_base = "http://localhost:8000/v1"
|
openai_api_base = "http://localhost:8000/v1"
|
||||||
chat_response = openai.ChatCompletion.create(
|
|
||||||
|
client = OpenAI(
|
||||||
|
api_key=openai_api_key,
|
||||||
|
base_url=openai_api_base,
|
||||||
|
)
|
||||||
|
|
||||||
|
chat_response = client.chat.completions.create(
|
||||||
model="facebook/opt-125m",
|
model="facebook/opt-125m",
|
||||||
messages=[
|
messages=[
|
||||||
{"role": "system", "content": "You are a helpful assistant."},
|
{"role": "system", "content": "You are a helpful assistant."},
|
||||||
|
|||||||
@ -1,18 +1,19 @@
|
|||||||
import openai
|
from openai import OpenAI
|
||||||
|
|
||||||
# Modify OpenAI's API key and API base to use vLLM's API server.
|
# Modify OpenAI's API key and API base to use vLLM's API server.
|
||||||
openai.api_key = "EMPTY"
|
openai_api_key = "EMPTY"
|
||||||
openai.api_base = "http://localhost:8000/v1"
|
openai_api_base = "http://localhost:8000/v1"
|
||||||
|
|
||||||
# List models API
|
client = OpenAI(
|
||||||
models = openai.Model.list()
|
# defaults to os.environ.get("OPENAI_API_KEY")
|
||||||
print("Models:", models)
|
api_key=openai_api_key,
|
||||||
|
base_url=openai_api_base,
|
||||||
|
)
|
||||||
|
|
||||||
model = models["data"][0]["id"]
|
models = client.models.list()
|
||||||
|
model = models.data[0].id
|
||||||
|
|
||||||
# Chat completion API
|
chat_completion = client.chat.completions.create(
|
||||||
chat_completion = openai.ChatCompletion.create(
|
|
||||||
model=model,
|
|
||||||
messages=[{
|
messages=[{
|
||||||
"role": "system",
|
"role": "system",
|
||||||
"content": "You are a helpful assistant."
|
"content": "You are a helpful assistant."
|
||||||
@ -27,7 +28,10 @@ chat_completion = openai.ChatCompletion.create(
|
|||||||
}, {
|
}, {
|
||||||
"role": "user",
|
"role": "user",
|
||||||
"content": "Where was it played?"
|
"content": "Where was it played?"
|
||||||
}])
|
}],
|
||||||
|
model=model,
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
print("Chat completion results:")
|
print("Chat completion results:")
|
||||||
print(chat_completion)
|
print(chat_completion)
|
||||||
|
|||||||
@ -1,24 +1,28 @@
|
|||||||
import openai
|
from openai import OpenAI
|
||||||
|
|
||||||
# Modify OpenAI's API key and API base to use vLLM's API server.
|
# Modify OpenAI's API key and API base to use vLLM's API server.
|
||||||
openai.api_key = "EMPTY"
|
openai_api_key = "EMPTY"
|
||||||
openai.api_base = "http://localhost:8000/v1"
|
openai_api_base = "http://localhost:8000/v1"
|
||||||
|
|
||||||
# List models API
|
client = OpenAI(
|
||||||
models = openai.Model.list()
|
# defaults to os.environ.get("OPENAI_API_KEY")
|
||||||
print("Models:", models)
|
api_key=openai_api_key,
|
||||||
|
base_url=openai_api_base,
|
||||||
|
)
|
||||||
|
|
||||||
model = models["data"][0]["id"]
|
models = client.models.list()
|
||||||
|
model = models.data[0].id
|
||||||
|
|
||||||
# Completion API
|
# Completion API
|
||||||
stream = False
|
stream = False
|
||||||
completion = openai.Completion.create(
|
completion = client.completions.create(
|
||||||
model=model,
|
model=model,
|
||||||
prompt="A robot may not injure a human being",
|
prompt="A robot may not injure a human being",
|
||||||
echo=False,
|
echo=False,
|
||||||
n=2,
|
n=2,
|
||||||
stream=stream,
|
stream=stream,
|
||||||
logprobs=3)
|
logprobs=3
|
||||||
|
)
|
||||||
|
|
||||||
print("Completion results:")
|
print("Completion results:")
|
||||||
if stream:
|
if stream:
|
||||||
|
|||||||
Loading…
x
Reference in New Issue
Block a user