Skip to content

Low GPU usage of quantized Mixtral 8x22B for prompt processing on Metal #6642

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Closed
beebopkim opened this issue Apr 12, 2024 · 2 comments
Closed

Comments

@beebopkim
Copy link

beebopkim commented Apr 12, 2024

My computer is M1 Max Mac Studio with 32 Cores of GPU with 64 GB of RAM. macOS version is Sonoma 14.4.1.

I run llama-bench from commit 4cc120c and it shows low GPU usage for prompt processing. Of course, inferences on main and server show same low GPU usages.

Screenshot 2024-04-13 at 12 40 32 AM-2

In the above image, I run benchmark for IQ2_XXS, IQ2_XS, IQ2_S, IQ2_M, and Q2_K_S but IQ1_S and IQ1_M from https://huggingface.co/MaziyarPanahi/Mixtral-8x22B-v0.1-GGUF will show same low GPU usage.

@stefanvarunix
Copy link

#6740

@github-actions github-actions bot added the stale label May 20, 2024
Copy link
Contributor

github-actions bot commented Jun 4, 2024

This issue was closed because it has been inactive for 14 days since being marked as stale.

@github-actions github-actions bot closed this as completed Jun 4, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

No branches or pull requests

2 participants