mirror of
https://github.com/vllm-project/vllm.git
synced 2025-10-20 14:53:52 +08:00
[Misc] Bump transformers version (#3592)
This commit is contained in:
@ -7,9 +7,9 @@ ray >= 2.9
|
||||
sentencepiece # Required for LLaMA tokenizer.
|
||||
numpy
|
||||
tokenizers>=0.15.0
|
||||
transformers >= 4.39.0 # Required for StarCoder2.
|
||||
transformers >= 4.39.1 # Required for StarCoder2 & Llava.
|
||||
fastapi
|
||||
uvicorn[standard]
|
||||
pydantic >= 2.0 # Required for OpenAI server.
|
||||
prometheus_client >= 0.18.0
|
||||
outlines == 0.0.34
|
||||
outlines == 0.0.34
|
||||
|
@ -5,7 +5,7 @@ ray >= 2.9
|
||||
sentencepiece # Required for LLaMA tokenizer.
|
||||
numpy
|
||||
torch == 2.1.2
|
||||
transformers >= 4.39.0 # Required for StarCoder2.
|
||||
transformers >= 4.39.1 # Required for StarCoder2 & Llava.
|
||||
xformers == 0.0.23.post1 # Required for CUDA 12.1.
|
||||
fastapi
|
||||
uvicorn[standard]
|
||||
|
Reference in New Issue
Block a user