Upgrade transformers to 5.x and other dependencies#1285
Upgrade transformers to 5.x and other dependencies#1285dxqb wants to merge 8 commits intoNerogar:masterfrom
Conversation
|
updating to CUDA 13 might be too early: |
|
CUDA 12.8 instead. Even many RTX 5090 clouds are still on drivers not supported by CUDA 13 |
This comment was marked as resolved.
This comment was marked as resolved.
This comment was marked as resolved.
This comment was marked as resolved.
|
Self note: Tested and the scalene upgrade breaks on windows. Need to look into it |
or downgrade to that version, but leave a comment that scalene must not be upgraded. which version? |
|
Unfortunately have to revert. Findings:
|
|
torch 2.10 alone first #1266 because this PR requires some more work |
|
On ROCm: It's only the first launch for any kernel that's slow on 2.10. It's actually a plus since most things are compiled on the fly, meaning the binary is actually smaller. So let's keep ROCm. |
|
when we do, #1346 can be reverted |
We're using only the parts of Scalene that didn't |
There are some major upgrades of dependencies:
CUDA 13.0This PR upgrades all dependencies to their latest version (except ROCM).
huggingface_hub was held back by transformers so far. With transformers 5.0, huggingface_hub gets a major upgrade.
This is why I've enabled XET in this PR. Previous problems with XET might not apply anymore - to be tested.
If XET works well, #770 could be reverted because you don't need to download multiple files anymore to get full speed.
Would close #949