-
-
Notifications
You must be signed in to change notification settings - Fork 10.6k
[Misc] Update to Transformers 4.48 #12120
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
Signed-off-by: Tyler Michael Smith <[email protected]>
👋 Hi! Thank you for contributing to the vLLM project. Once the PR is approved and ready to go, your PR reviewer(s) can run CI to test the changes comprehensively before merging. To run CI, PR reviewers can do one of these:
🚀 |
FYI @mgoin, @dsikka, @robertgshaw2-redhat on the neuralmagic/Llama-3.2-1B-quantized.w8a8 issue |
The transformers 4.48 release has the required change in how Automodel is expected to handle compressed vs frozen models. The model in question has the frozen state, which really should be compressed. Likely just outdated I'll look into updating the model |
The custom model in question that is causing Language Model Test to fail is this one in particular. What can we do about this since this code is only modifiable but the its maintainers?.
|
The same issue has appeared in a few other models as a result of the release e.g Deepseek |
I put up a PR to fix those ch
I have a PR up to fix it just for MiniCMP here: https://huggingface.co/openbmb/MiniCPM3-4B/discussions/39 But alternatively if there are multiple models with the same issue, I think it would be better to fix via huggingface/transformers#35734 -- @dsikka do you have links to other models with the same issue? |
FYI the model has been updated and the tests now pass (at least locally) |
The basic correctness tests pass for me if using this: huggingface/transformers#35681. For now we can wait until it gets merged, or skip the tests until the new point release of transformers |
Added the ready label to see if the Multi GPU test is intermittent |
Looks like the multi gpu tests are passing so we should be good to go once huggingface/transformers#35681 lands. I'm inclined to wait for a transformers point release but lmk if you disagree |
@tlrmchlsmth thanks! Im absolutely on board with waiting for the point release |
huggingface/transformers#35681 is now merged, so that's one less error as soon as they make a point release after it! |
It still happens in |
Patch release 4.48.2 is out btw ! |
Signed-off-by: Tyler Michael Smith <[email protected]>
Signed-off-by: Tyler Michael Smith <[email protected]>
@tlrmchlsmth seems like few tests failed, its wierd because some of these tests (e.g., Open AI, TPU test) passed before when I first tried to perform the upgrade to 4.48.0 in my other PR #10909. Wonder if some of the are intermittent |
The CI for TPUs is broken due to disks being full. Not sure about the others - I'm rerunning the failed tests |
Closing as superseded by #12599 |
Update Transformers to 4.48. Split off from #10909 to isolate any 4.48-related changes here and for easier debugging.
From @fabianlim, we have the following open issues: