mirror of
https://git.datalinker.icu/vllm-project/vllm.git
synced 2025-12-14 04:04:57 +08:00
Upgrade transformers version to 4.36.0 (#2046)
This commit is contained in:
parent
f3e024bece
commit
cb3f30c600
@ -10,7 +10,7 @@ numpy
|
|||||||
tokenizers>=0.15.0
|
tokenizers>=0.15.0
|
||||||
huggingface_hub<0.18,>=0.16.4
|
huggingface_hub<0.18,>=0.16.4
|
||||||
einops # Required for phi-1_5
|
einops # Required for phi-1_5
|
||||||
transformers >= 4.34.0 # Required for Mistral.
|
transformers >= 4.36.0 # Required for Mixtral.
|
||||||
fastapi
|
fastapi
|
||||||
uvicorn[standard]
|
uvicorn[standard]
|
||||||
pydantic == 1.10.13 # Required for OpenAI server.
|
pydantic == 1.10.13 # Required for OpenAI server.
|
||||||
|
|||||||
@ -7,7 +7,7 @@ sentencepiece # Required for LLaMA tokenizer.
|
|||||||
numpy
|
numpy
|
||||||
einops # Required for phi-1_5
|
einops # Required for phi-1_5
|
||||||
torch >= 2.1.1
|
torch >= 2.1.1
|
||||||
transformers >= 4.34.0 # Required for Mistral.
|
transformers >= 4.36.0 # Required for Mixtral.
|
||||||
xformers >= 0.0.23 # Required for CUDA 12.1.
|
xformers >= 0.0.23 # Required for CUDA 12.1.
|
||||||
fastapi
|
fastapi
|
||||||
uvicorn[standard]
|
uvicorn[standard]
|
||||||
|
|||||||
@ -29,7 +29,7 @@ import torch
|
|||||||
import torch.nn.functional as F
|
import torch.nn.functional as F
|
||||||
|
|
||||||
from torch import nn
|
from torch import nn
|
||||||
from transformers import MistralConfig
|
from transformers import MixtralConfig
|
||||||
|
|
||||||
try:
|
try:
|
||||||
import megablocks.ops as ops
|
import megablocks.ops as ops
|
||||||
@ -395,7 +395,7 @@ class MixtralDecoderLayer(nn.Module):
|
|||||||
|
|
||||||
def __init__(
|
def __init__(
|
||||||
self,
|
self,
|
||||||
config: MistralConfig,
|
config: MixtralConfig,
|
||||||
) -> None:
|
) -> None:
|
||||||
super().__init__()
|
super().__init__()
|
||||||
self.hidden_size = config.hidden_size
|
self.hidden_size = config.hidden_size
|
||||||
@ -443,7 +443,7 @@ class MixtralForCausalLM(nn.Module):
|
|||||||
|
|
||||||
def __init__(
|
def __init__(
|
||||||
self,
|
self,
|
||||||
config: MistralConfig,
|
config: MixtralConfig,
|
||||||
linear_method: Optional[LinearMethodBase] = None,
|
linear_method: Optional[LinearMethodBase] = None,
|
||||||
) -> None:
|
) -> None:
|
||||||
super().__init__()
|
super().__init__()
|
||||||
|
|||||||
Loading…
x
Reference in New Issue
Block a user