Skip to content

Bump mlx-lm/mlx-vlm deps and add Qwen3.5-0.8B smoke test #174

Merged
LxYuan0420 merged 1 commit intovllm-project:mainfrom
ricky-chaoju:deps/bump-mlx-deps
Mar 19, 2026
Merged

Bump mlx-lm/mlx-vlm deps and add Qwen3.5-0.8B smoke test #174
LxYuan0420 merged 1 commit intovllm-project:mainfrom
ricky-chaoju:deps/bump-mlx-deps

Conversation

@ricky-chaoju
Copy link
Contributor

@ricky-chaoju ricky-chaoju commented Mar 18, 2026

Summary

Follow-up to #169 (transformers>=5.0.0 upgrade).

  • Bump mlx>=0.31.0, mlx-lm>=0.31.0, mlx-vlm>=0.4.0
  • Add tests/test_qwen35_smoke.py: golden token comparison for Qwen/Qwen3.5-0.8B (greedy decoding, 5/5 passed)

Qwen3.5 uses the qwen3_5 architecture which requires transformers>=5.0.0 and mlx-lm>=0.30.0. This proves the upgraded dependency stack works end-to-end on Metal.

Test

test_qwen35_smoke

截圖 2026-03-18 下午5 09 56

test_paged_deterministic

截圖 2026-03-18 下午5 10 22

@LxYuan0420
Copy link
Collaborator

Merged #169 now, sorry for the delay. LGTM and just needs a rebase since the transformers bump commit (b2fa307) is already on main now.

Signed-off-by: Chao-Ju Chen <ricky.chen@infinirc.com>
@ricky-chaoju
Copy link
Contributor Author

done! Thanks @LxYuan0420

@ricky-chaoju
Copy link
Contributor Author

@LxYuan0420 Please check #176 first before merging this one. Thanks!

@WindChimeRan
Copy link
Collaborator

@ricky-chaoju Thanks! This PR will give us groundtruth labels and unblock the flash-linear-attn for the paged path!

@LxYuan0420 LxYuan0420 merged commit 3c683ea into vllm-project:main Mar 19, 2026
7 of 13 checks passed
@ricky-chaoju ricky-chaoju deleted the deps/bump-mlx-deps branch March 19, 2026 02:39
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants