We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 9938bba commit 93dfc82Copy full SHA for 93dfc82
vllm/model_executor/layers/batch_invariant.py
@@ -11,6 +11,7 @@
11
import vllm.envs as envs
12
from vllm.logger import init_logger
13
from vllm.triton_utils import tl, triton
14
+from vllm.utils.torch_utils import is_torch_equal_or_newer
15
16
logger = init_logger(__name__)
17
@@ -741,8 +742,6 @@ def disable_batch_invariant_mode():
741
742
_batch_invariant_MODE = False
743
_batch_invariant_LIB = None
744
- torch.backends.cuda.preferred_blas_library(backend="default")
745
-
746
747
@contextlib.contextmanager
748
def set_batch_invariant_mode(enabled: bool = True):
0 commit comments