Skip to content
Closed
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
2 changes: 1 addition & 1 deletion requirements/common.txt
Original file line number Diff line number Diff line change
Expand Up @@ -7,7 +7,7 @@ requests >= 2.26.0
tqdm
blake3
py-cpuinfo
transformers >= 4.56.0, < 5
transformers >= 4.56.0, <= 5.2.0
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

critical

Relaxing the upper bound for the transformers library from < 5 to <= 5.2.0 is a high-risk change as it crosses a major version boundary. Major versions of libraries often introduce breaking changes that could negatively impact vLLM's functionality and stability.

The pull request description does not provide a test plan or any evidence that vLLM has been thoroughly tested against transformers versions between 5.0.0 and 5.2.0. Without this validation, we cannot be confident that this change will not introduce regressions.

Please add a comprehensive test plan and the corresponding results to demonstrate that vLLM remains fully functional with the newly allowed versions of transformers. If this cannot be provided, it would be safer to revert this change to maintain the project's stability.

Copy link
Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Thanks for the detailed feedback — agreed on the risk of crossing a major-version boundary.
I will add a concrete validation matrix and results for transformers==5.0.0, 5.1.0, and 5.2.0, including:

vLLM install/import checks
offline generation smoke test
OpenAI-compatible server startup + one request
tokenizer/chat-template sanity checks
If any incompatibility is found, I will narrow the upper bound or revert this change.

tokenizers >= 0.21.1 # Required for fast incremental detokenization.
protobuf >= 5.29.6, !=6.30.*, !=6.31.*, !=6.32.*, !=6.33.0.*, !=6.33.1.*, !=6.33.2.*, !=6.33.3.*, !=6.33.4.* # Required by LlamaTokenizer, gRPC. CVE-2026-0994
fastapi[standard] >= 0.115.0 # Required by FastAPI's form models in the OpenAI API server's audio transcriptions endpoint.
Expand Down