You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Add batched Llama model definition using vLLM paged attention (mlc-ai#1134)
* Add batched Llama model with vllm paged attention
* update core.py
* doc
* minor
* add e2e test
* mv file
* clean
* Check if TVM has been built with USE_VLLM
* update BuildArgs docstring
0 commit comments