-
Notifications
You must be signed in to change notification settings - Fork 6.2k
[Revision] Don't recommend using revision #1764
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
The documentation is not available anymore as the PR was closed or merged. |
Err, wait, but don't all the flagship Stable Diffusion models serve fp32 weights on their default branch? The bandwidth and storage advantages of fp16 models are significant. Literally half the space. Pretty sure it's a significant factor in model loading speed too. At least for people using fp16, which I think just about everyone using CUDA for Stable Diffusion is. I very much hope you continue to provide these half-precision model weights (for the models for which they are viable, anyway), and I encourage you to not underestimate how much people on commodity hardware will appreciate the lower resource requirements! That said, you have identified a usability issue here: there's no API to specify your architecture when asking the hub for a model, there's not a well-established convention for which branches or subfolders people use for this, and I didn't have an easy go of it when trying to figure out how to detect if a repo has a fp16 branch or not. |
Agree with @keturn, the download, disk space and load speed of fp16 versions are significant. It's not a blocker or anything, but it would make me sad if fp16 stopped being provided. (Although caveat: it's easy to accidentally download both fp16 and fp32 models, thereby making two of those worse :) ) @keturn although I don't do it yet, you should be able to use |
I agree that we should try to download the fp16 versions if they are available and the user specifies a
We don't currently have a general solution to solve this, just some pieces to deal with safetensors vs PyTorch files. But in my opinion it's something we should work on soon. |
Always one step ahead @keturn 😊 We're currently looking into a better way to support "fp16" directly for download. Our current thoughts are:
This will be a bigger refactor/deprecation PR, but it should be better because:
We will soon open a bigger PR for this to discuss with the community. Currently the idea is to create the following naming scheme:
Wdyt about this @keturn @hafriedlander ? (cc @pcuenca) |
We should not recommend using the "fp16" revision anymore since most model checkpoints don't have such a branch and also because it's quite an unintuitive behavior.