-
-
Notifications
You must be signed in to change notification settings - Fork 10.2k
Description
pip install vllm (0.6.3) will force a reinstallation of the CPU version torch and replace cuda torch on windows. pip install vllm(0.6.3)将强制重新安装CPU版本的torch并在Windows上替换cuda torch。
I don't quite get what you mean, how can you have different versions of torch for CPU and GPU at the same time?我不太明白你的意思,你怎么能有不同版本的火炬CPU和GPU在同一时间?
only cuda torch
pip install vllm --no-deps Collecting vllm Using cached vllm-0.6.3.post1.tar.gz (2.7 MB) Installing build dependencies ... error error: subprocess-exited-with-error × pip subprocess to install build dependencies did not run successfully. │ exit code: 2 ╰─> [86 lines of output] Collecting cmake>=3.26 Using cached cmake-3.30.5-py3-none-win_amd64.whl.metadata (6.4 kB) Collecting ninja Using cached ninja-1.11.1.1-py2.py3-none-win_amd64.whl.metadata (5.4 kB) Collecting packaging Using cached packaging-24.1-py3-none-any.whl.metadata (3.2 kB) Collecting setuptools>=61 Using cached setuptools-75.2.0-py3-none-any.whl.metadata (6.9 kB) Collecting setuptools-scm>=8.0 Using cached setuptools_scm-8.1.0-py3-none-any.whl.metadata (6.6 kB) Collecting torch==2.4.0 Using cached torch-2.4.0-cp310-cp310-win_amd64.whl.metadata (27 kB) Collecting wheel Using cached wheel-0.44.0-py3-none-any.whl.metadata (2.3 kB) Collecting jinja2 Using cached jinja2-3.1.4-py3-none-any.whl.metadata (2.6 kB) Collecting filelock (from torch==2.4.0) Using cached filelock-3.16.1-py3-none-any.whl.metadata (2.9 kB) Collecting typing-extensions>=4.8.0 (from torch==2.4.0) Using cached typing_extensions-4.12.2-py3-none-any.whl.metadata (3.0 kB) Collecting sympy (from torch==2.4.0) Using cached sympy-1.13.3-py3-none-any.whl.metadata (12 kB) Collecting networkx (from torch==2.4.0) Using cached networkx-3.4.2-py3-none-any.whl.metadata (6.3 kB) Collecting fsspec (from torch==2.4.0) Using cached fsspec-2024.10.0-py3-none-any.whl.metadata (11 kB) Collecting tomli>=1 (from setuptools-scm>=8.0) Using cached tomli-2.0.2-py3-none-any.whl.metadata (10.0 kB) Collecting MarkupSafe>=2.0 (from jinja2) Using cached MarkupSafe-3.0.2-cp310-cp310-win_amd64.whl.metadata (4.1 kB ) Collecting mpmath<1.4,>=1.1.0 (from sympy->torch==2.4.0) Using cached mpmath-1.3.0-py3-none-any.whl.metadata (8.6 kB) Downloading torch-2.4.0-cp310-cp310-win_amd64.whl (197.9 MB) 3.9/197.9 MB 21.3 kB/s eta 2:3 1:31 ERROR: Exception: Traceback (most recent call last): File "D:\my\env\python3.10.10\Lib\site-packages\pip\_vendor\urllib3\resp onse.py", line 438, in _error_catcher yield File "D:\my\env\python3.10.10\Lib\site-packages\pip\_vendor\urllib3\resp onse.py", line 561, in read data = self._fp_read(amt) if not fp_closed else b"" File "D:\my\env\python3.10.10\Lib\site-packages\pip\_vendor\urllib3\resp onse.py", line 527, in _fp_read return self._fp.read(amt) if amt is not None else self._fp.read() File "D:\my\env\python3.10.10\Lib\site-packages\pip\_vendor\cachecontrol \filewrapper.py", line 98, in read data: bytes = self.__fp.read(amt) File "D:\my\env\python3.10.10\lib\http\client.py", line 465, in read s = self.fp.read(amt) File "D:\my\env\python3.10.10\lib\socket.py", line 705, in readinto return self._sock.recv_into(b) File "D:\my\env\python3.10.10\lib\ssl.py", line 1274, in recv_into return self.read(nbytes, buffer) File "D:\my\env\python3.10.10\lib\ssl.py", line 1130, in read return self._sslobj.read(len, buffer) TimeoutError: The read operation timed out During handling of the above exception, another exception occurred: Traceback (most recent call last): File "D:\my\env\python3.10.10\Lib\site-packages\pip\_internal\cli\base_c ommand.py", line 105, in _run_wrapper status = _inner_run() File "D:\my\env\python3.10.10\Lib\site-packages\pip\_internal\cli\base_c ommand.py", line 96, in _inner_run return self.run(options, args) File "D:\my\env\python3.10.10\Lib\site-packages\pip\_internal\cli\req_co mmand.py", line 67, in wrapper return func(self, options, args) File "D:\my\env\python3.10.10\Lib\site-packages\pip\_internal\commands\i nstall.py", line 379, in run requirement_set = resolver.resolve( File "D:\my\env\python3.10.10\Lib\site-packages\pip\_internal\resolution \resolvelib\resolver.py", line 179, in resolve self.factory.preparer.prepare_linked_requirements_more(reqs) File "D:\my\env\python3.10.10\Lib\site-packages\pip\_internal\operations \prepare.py", line 554, in prepare_linked_requirements_more self._complete_partial_requirements( File "D:\my\env\python3.10.10\Lib\site-packages\pip\_internal\operations \prepare.py", line 469, in _complete_partial_requirements for link, (filepath, _) in batch_download: File "D:\my\env\python3.10.10\Lib\site-packages\pip\_internal\network\do wnload.py", line 184, in __call__ for chunk in chunks: File "D:\my\env\python3.10.10\Lib\site-packages\pip\_internal\cli\progre ss_bars.py", line 55, in _rich_progress_bar for chunk in iterable: File "D:\my\env\python3.10.10\Lib\site-packages\pip\_internal\network\ut ils.py", line 65, in response_chunks for chunk in response.raw.stream( File "D:\my\env\python3.10.10\Lib\site-packages\pip\_vendor\urllib3\resp onse.py", line 622, in stream data = self.read(amt=amt, decode_content=decode_content) File "D:\my\env\python3.10.10\Lib\site-packages\pip\_vendor\urllib3\resp onse.py", line 560, in read with self._error_catcher(): File "D:\my\env\python3.10.10\lib\contextlib.py", line 153, in __exit__ self.gen.throw(typ, value, traceback) File "D:\my\env\python3.10.10\Lib\site-packages\pip\_vendor\urllib3\resp onse.py", line 443, in _error_catcher raise ReadTimeoutError(self._pool, None, "Read timed out.") pip._vendor.urllib3.exceptions.ReadTimeoutError: HTTPSConnectionPool(host= 'files.pythonhosted.org', port=443): Read timed out. [end of output] note: This error originates from a subprocess, and is likely not a problem wit h pip. error: subprocess-exited-with-error × pip subprocess to install build dependencies did not run successfully. │ exit code: 2 ╰─> See above for output.
If you internet is not good. You are so lucky. Because it will fail during the process of forcibly replacing CUDA torch with CPU. If you have a good internet connection. So things will become very bad. Your torch will transition from CUDA to a lower version CPU. And pip install vllm --no-deps or pip install vllm has same issue
What is your original version of pytorch?
Originally posted by @DarkLight1337 in #4194 (comment)
pip show torch
Name: torch
Version: 2.5.0+cu124
Summary: Tensors and Dynamic neural networks in Python with strong GPU acceleration
Home-page: https://pytorch.org/
Author: PyTorch Team
Author-email: [email protected]
License: BSD-3-Clause
Location: d:\my\env\python3.10.10\lib\site-packages
Requires: filelock, fsspec, jinja2, networkx, sympy, typing-extensions
Required-by: accelerate, auto_gptq, bitsandbytes, compressed-tensors, encodec, flash_attn, optimum, peft, stable-baselines3, timm, t
orchaudio, torchvision, trl, vector-quantize-pytorch, vocos