Closed ramyaprabhu-alt closed 7 months ago
Uninstall the package called transformer-engine by using the command pip uninstall transformer-engine
. Install transformers
and accelerate
if not already installed. Then it should work.
Uninstall the package called transformer-engine by using the command
pip uninstall transformer-engine
. Installtransformers
andaccelerate
if not already installed. Then it should work.
I tried this but didn't work
Please post the vllm version and then the steps to reproduce this. Which model are you using? Cuda version on the system and the docker (if you use it).
I am also experiencing this issue.
env: vllm-0.2.2 , cuda 12.2 , no docker. Local build of vllm from source, then running examples/offline_inference.py
.
i also experience vllm==0.2.5 / accelerate==0.27.2 / transformers==4.37.2
Any idea? I am also facing this issue Docker image: NGC docker container 23.10-py3 too.
same problem, I am running on Kaggle.com
Successfully installed aioprometheus-23.12.0 cupy-cuda12x-12.1.0 pynvml-11.5.0 quantile-python-1.1 transformers-4.38.1 triton-2.2.0 vllm-0.3.2 xformers-0.0.23.post1
It seems to be PyTorch isn't working with Cuda-12.2 on NGC-PyTorch 23.10-py3. I fix this issue on NGC-PyTorch 23.07-py3 and reinstall pytorch with v2.1.2
same issue. anyone able to fix?
cuda: 12.0.1 torch : 2.2.1 transformers: 4.38.2 vllm: 0.3.2 accelerate: 0.22.0
UPDATE: solved by downgrading torch to 2.1.2
I'm hitting this same problem :(
@sudarshan-kamath , pip uninstall transformer-engine
also did not work for me
@RylanSchaeffer can you try to installing a matching version of pytorch with vllm?
e.g. vllm 0.3.3 with pytorch 2.1 .
Versions:
>>> import torch
>>> torch.__version__
'2.1.2.post301'
>>> import vllm
>>> vllm.__version__
'0.3.3'
Error:
ImportError: /lfs/ampere8/0/rschaef/miniconda3/envs/model_collapse_env/lib/python3.10/site-packages/vllm/_C.cpython-310-x86_64-linux-gnu.so: undefined symbol: _ZN3c106detail23torchInternalAssertFailEPKcS2_jS2_RKSs
@youkaichao , here's a script I'm using to debug:
from vllm import LLM, SamplingParams
sampling_params = SamplingParams(max_tokens=50, temperature=1.0)
llm = LLM(
model=f"gpt2",
tensor_parallel_size=4,
# load_format="pt",
)
outputs = llm.generate(
["My name is Lewis and I like to", "Today is a good day to", "John has"],
sampling_params,
)
for output in outputs:
print(output.outputs[0].text)
print("=========")
2.1.2.post301
Are you using a custom built version of pytorch?
vLLM is compiled against officially released pytorch. And there is no binary compatibility promise across pytorch versions.
You can try to build vllm yourself: https://docs.vllm.ai/en/latest/getting_started/installation.html#build-from-source .
Are you using a custom built version of pytorch?
No, I installed using the default command from pytorch itself (conda install pytorch torchvision torchaudio pytorch-cuda=12.1 -c pytorch -c nvidia
)
I am now trying the following: I deleted my conda environment, added vllm
as a git submodules and am now trying pip install -e .
This is the error I received:
Building wheels for collected packages: vllm
Building editable for vllm (pyproject.toml) ... error
error: subprocess-exited-with-error
× Building editable for vllm (pyproject.toml) did not run successfully.
│ exit code: 1
╰─> [145 lines of output]
/tmp/user/21163/pip-build-env-a245et9d/overlay/lib/python3.10/site-packages/torch/nn/modules/transformer.py:20: UserWarning: Failed to initialize NumPy: No module named 'numpy' (Triggered internally at ../torch/csrc/utils/tensor_numpy.cpp:84.)
device: torch.device = torch.device(torch._C._get_default_device()), # torch.device('cpu'),
running editable_wheel
creating /tmp/user/21163/pip-wheel-zf61weqe/.tmp-id2wfdg6/vllm.egg-info
writing /tmp/user/21163/pip-wheel-zf61weqe/.tmp-id2wfdg6/vllm.egg-info/PKG-INFO
writing dependency_links to /tmp/user/21163/pip-wheel-zf61weqe/.tmp-id2wfdg6/vllm.egg-info/dependency_links.txt
writing requirements to /tmp/user/21163/pip-wheel-zf61weqe/.tmp-id2wfdg6/vllm.egg-info/requires.txt
writing top-level names to /tmp/user/21163/pip-wheel-zf61weqe/.tmp-id2wfdg6/vllm.egg-info/top_level.txt
writing manifest file '/tmp/user/21163/pip-wheel-zf61weqe/.tmp-id2wfdg6/vllm.egg-info/SOURCES.txt'
reading manifest file '/tmp/user/21163/pip-wheel-zf61weqe/.tmp-id2wfdg6/vllm.egg-info/SOURCES.txt'
reading manifest template 'MANIFEST.in'
adding license file 'LICENSE'
writing manifest file '/tmp/user/21163/pip-wheel-zf61weqe/.tmp-id2wfdg6/vllm.egg-info/SOURCES.txt'
creating '/tmp/user/21163/pip-wheel-zf61weqe/.tmp-id2wfdg6/vllm-0.3.3+cu122.dist-info'
creating /tmp/user/21163/pip-wheel-zf61weqe/.tmp-id2wfdg6/vllm-0.3.3+cu122.dist-info/WHEEL
running build_py
running build_ext
-- The CXX compiler identification is GNU 9.4.0
-- Detecting CXX compiler ABI info
-- Detecting CXX compiler ABI info - done
-- Check for working CXX compiler: /usr/bin/c++ - skipped
-- Detecting CXX compile features
-- Detecting CXX compile features - done
-- Build type: RelWithDebInfo
-- Found Python: /lfs/ampere8/0/rschaef/miniconda3/envs/model_collapse_env/bin/python3.10 (found version "3.10.14") found components: Interpreter Development.Module
-- Found python matching: /lfs/ampere8/0/rschaef/miniconda3/envs/model_collapse_env/bin/python3.10.
-- Found CUDA: /usr/local/cuda (found version "12.2")
-- The CUDA compiler identification is unknown
CMake Error at /tmp/user/21163/pip-build-env-a245et9d/overlay/lib/python3.10/site-packages/cmake/data/share/cmake-3.28/Modules/CMakeDetermineCUDACompiler.cmake:270 (message):
Failed to detect a default CUDA architecture.
Compiler output:
Call Stack (most recent call first):
/tmp/user/21163/pip-build-env-a245et9d/overlay/lib/python3.10/site-packages/torch/share/cmake/Caffe2/public/cuda.cmake:47 (enable_language)
/tmp/user/21163/pip-build-env-a245et9d/overlay/lib/python3.10/site-packages/torch/share/cmake/Caffe2/Caffe2Config.cmake:87 (include)
/tmp/user/21163/pip-build-env-a245et9d/overlay/lib/python3.10/site-packages/torch/share/cmake/Torch/TorchConfig.cmake:68 (find_package)
CMakeLists.txt:64 (find_package)
-- Configuring incomplete, errors occurred!
Traceback (most recent call last):
File "/tmp/user/21163/pip-build-env-a245et9d/overlay/lib/python3.10/site-packages/setuptools/command/editable_wheel.py", line 155, in run
self._create_wheel_file(bdist_wheel)
File "/tmp/user/21163/pip-build-env-a245et9d/overlay/lib/python3.10/site-packages/setuptools/command/editable_wheel.py", line 357, in _create_wheel_file
files, mapping = self._run_build_commands(dist_name, unpacked, lib, tmp)
File "/tmp/user/21163/pip-build-env-a245et9d/overlay/lib/python3.10/site-packages/setuptools/command/editable_wheel.py", line 280, in _run_build_commands
self._run_build_subcommands()
File "/tmp/user/21163/pip-build-env-a245et9d/overlay/lib/python3.10/site-packages/setuptools/command/editable_wheel.py", line 307, in _run_build_subcommands
self.run_command(name)
File "/tmp/user/21163/pip-build-env-a245et9d/overlay/lib/python3.10/site-packages/setuptools/_distutils/cmd.py", line 318, in run_command
self.distribution.run_command(command)
File "/tmp/user/21163/pip-build-env-a245et9d/overlay/lib/python3.10/site-packages/setuptools/dist.py", line 967, in run_command
super().run_command(command)
File "/tmp/user/21163/pip-build-env-a245et9d/overlay/lib/python3.10/site-packages/setuptools/_distutils/dist.py", line 988, in run_command
cmd_obj.run()
File "/tmp/user/21163/pip-build-env-a245et9d/overlay/lib/python3.10/site-packages/setuptools/command/build_ext.py", line 91, in run
_build_ext.run(self)
File "/tmp/user/21163/pip-build-env-a245et9d/overlay/lib/python3.10/site-packages/setuptools/_distutils/command/build_ext.py", line 345, in run
self.build_extensions()
File "<string>", line 164, in build_extensions
File "<string>", line 147, in configure
File "/lfs/ampere8/0/rschaef/miniconda3/envs/model_collapse_env/lib/python3.10/subprocess.py", line 369, in check_call
raise CalledProcessError(retcode, cmd)
subprocess.CalledProcessError: Command '['cmake', '/lfs/ampere8/0/rschaef/KoyejoLab-Revisiting-Model-Collapse/submodules/vllm', '-G', 'Ninja', '-DCMAKE_BUILD_TYPE=RelWithDebInfo', '-DCMAKE_LIBRARY_OUTPUT_DIRECTORY=/tmp/user/21163/tmproow4us4.build-lib/vllm', '-DCMAKE_ARCHIVE_OUTPUT_DIRECTORY=/tmp/user/21163/tmpqy5i3vwd.build-temp', '-DVLLM_PYTHON_EXECUTABLE=/lfs/ampere8/0/rschaef/miniconda3/envs/model_collapse_env/bin/python3.10', '-DNVCC_THREADS=8', '-DCMAKE_JOB_POOL_COMPILE:STRING=compile', '-DCMAKE_JOB_POOLS:STRING=compile=64']' returned non-zero exit status 1.
/tmp/user/21163/pip-build-env-a245et9d/overlay/lib/python3.10/site-packages/setuptools/_distutils/dist.py:988: _DebuggingTips: Problem in editable installation.
!!
********************************************************************************
An error happened while installing `vllm` in editable mode.
The following steps are recommended to help debug this problem:
- Try to install the project normally, without using the editable mode.
Does the error still persist?
(If it does, try fixing the problem before attempting the editable mode).
- If you are using binary extensions, make sure you have all OS-level
dependencies installed (e.g. compilers, toolchains, binary libraries, ...).
- Try the latest version of setuptools (maybe the error was already fixed).
- If you (or your project dependencies) are using any setuptools extension
or customization, make sure they support the editable mode.
After following the steps above, if the problem still persists and
you think this is related to how setuptools handles editable installations,
please submit a reproducible example
(see https://stackoverflow.com/help/minimal-reproducible-example) to:
https://github.com/pypa/setuptools/issues
See https://setuptools.pypa.io/en/latest/userguide/development_mode.html for details.
********************************************************************************
!!
cmd_obj.run()
Traceback (most recent call last):
File "/lfs/ampere8/0/rschaef/miniconda3/envs/model_collapse_env/lib/python3.10/site-packages/pip/_vendor/pyproject_hooks/_in_process/_in_process.py", line 353, in <module>
main()
File "/lfs/ampere8/0/rschaef/miniconda3/envs/model_collapse_env/lib/python3.10/site-packages/pip/_vendor/pyproject_hooks/_in_process/_in_process.py", line 335, in main
json_out['return_val'] = hook(**hook_input['kwargs'])
File "/lfs/ampere8/0/rschaef/miniconda3/envs/model_collapse_env/lib/python3.10/site-packages/pip/_vendor/pyproject_hooks/_in_process/_in_process.py", line 273, in build_editable
return hook(wheel_directory, config_settings, metadata_directory)
File "/tmp/user/21163/pip-build-env-a245et9d/overlay/lib/python3.10/site-packages/setuptools/build_meta.py", line 443, in build_editable
return self._build_with_temp_dir(
File "/tmp/user/21163/pip-build-env-a245et9d/overlay/lib/python3.10/site-packages/setuptools/build_meta.py", line 395, in _build_with_temp_dir
self.run_setup()
File "/tmp/user/21163/pip-build-env-a245et9d/overlay/lib/python3.10/site-packages/setuptools/build_meta.py", line 311, in run_setup
exec(code, locals())
File "<string>", line 345, in <module>
File "/tmp/user/21163/pip-build-env-a245et9d/overlay/lib/python3.10/site-packages/setuptools/__init__.py", line 104, in setup
return distutils.core.setup(**attrs)
File "/tmp/user/21163/pip-build-env-a245et9d/overlay/lib/python3.10/site-packages/setuptools/_distutils/core.py", line 185, in setup
return run_commands(dist)
File "/tmp/user/21163/pip-build-env-a245et9d/overlay/lib/python3.10/site-packages/setuptools/_distutils/core.py", line 201, in run_commands
dist.run_commands()
File "/tmp/user/21163/pip-build-env-a245et9d/overlay/lib/python3.10/site-packages/setuptools/_distutils/dist.py", line 969, in run_commands
self.run_command(cmd)
File "/tmp/user/21163/pip-build-env-a245et9d/overlay/lib/python3.10/site-packages/setuptools/dist.py", line 967, in run_command
super().run_command(command)
File "/tmp/user/21163/pip-build-env-a245et9d/overlay/lib/python3.10/site-packages/setuptools/_distutils/dist.py", line 988, in run_command
cmd_obj.run()
File "/tmp/user/21163/pip-build-env-a245et9d/overlay/lib/python3.10/site-packages/setuptools/command/editable_wheel.py", line 155, in run
self._create_wheel_file(bdist_wheel)
File "/tmp/user/21163/pip-build-env-a245et9d/overlay/lib/python3.10/site-packages/setuptools/command/editable_wheel.py", line 357, in _create_wheel_file
files, mapping = self._run_build_commands(dist_name, unpacked, lib, tmp)
File "/tmp/user/21163/pip-build-env-a245et9d/overlay/lib/python3.10/site-packages/setuptools/command/editable_wheel.py", line 280, in _run_build_commands
self._run_build_subcommands()
File "/tmp/user/21163/pip-build-env-a245et9d/overlay/lib/python3.10/site-packages/setuptools/command/editable_wheel.py", line 307, in _run_build_subcommands
self.run_command(name)
File "/tmp/user/21163/pip-build-env-a245et9d/overlay/lib/python3.10/site-packages/setuptools/_distutils/cmd.py", line 318, in run_command
self.distribution.run_command(command)
File "/tmp/user/21163/pip-build-env-a245et9d/overlay/lib/python3.10/site-packages/setuptools/dist.py", line 967, in run_command
super().run_command(command)
File "/tmp/user/21163/pip-build-env-a245et9d/overlay/lib/python3.10/site-packages/setuptools/_distutils/dist.py", line 988, in run_command
cmd_obj.run()
File "/tmp/user/21163/pip-build-env-a245et9d/overlay/lib/python3.10/site-packages/setuptools/command/build_ext.py", line 91, in run
_build_ext.run(self)
File "/tmp/user/21163/pip-build-env-a245et9d/overlay/lib/python3.10/site-packages/setuptools/_distutils/command/build_ext.py", line 345, in run
self.build_extensions()
File "<string>", line 164, in build_extensions
File "<string>", line 147, in configure
File "/lfs/ampere8/0/rschaef/miniconda3/envs/model_collapse_env/lib/python3.10/subprocess.py", line 369, in check_call
raise CalledProcessError(retcode, cmd)
subprocess.CalledProcessError: Command '['cmake', '/lfs/ampere8/0/rschaef/KoyejoLab-Revisiting-Model-Collapse/submodules/vllm', '-G', 'Ninja', '-DCMAKE_BUILD_TYPE=RelWithDebInfo', '-DCMAKE_LIBRARY_OUTPUT_DIRECTORY=/tmp/user/21163/tmproow4us4.build-lib/vllm', '-DCMAKE_ARCHIVE_OUTPUT_DIRECTORY=/tmp/user/21163/tmpqy5i3vwd.build-temp', '-DVLLM_PYTHON_EXECUTABLE=/lfs/ampere8/0/rschaef/miniconda3/envs/model_collapse_env/bin/python3.10', '-DNVCC_THREADS=8', '-DCMAKE_JOB_POOL_COMPILE:STRING=compile', '-DCMAKE_JOB_POOLS:STRING=compile=64']' returned non-zero exit status 1.
[end of output]
note: This error originates from a subprocess, and is likely not a problem with pip.
ERROR: Failed building editable for vllm
Failed to build vllm
ERROR: Could not build wheels for vllm, which is required to install pyproject.toml-based projects
I'm deleting my pip
and conda
cache and trying again
I purged my .cache/pip
and .cache/conda
. Trying pip install -e .
yielded the same error:
Building wheels for collected packages: vllm
Building editable for vllm (pyproject.toml) ... error
error: subprocess-exited-with-error
× Building editable for vllm (pyproject.toml) did not run successfully.
│ exit code: 1
╰─> [145 lines of output]
/tmp/user/21163/pip-build-env-ygusfc74/overlay/lib/python3.10/site-packages/torch/nn/modules/transformer.py:20: UserWarning: Failed to initialize NumPy: No module named 'numpy' (Triggered internally at ../torch/csrc/utils/tensor_numpy.cpp:84.)
device: torch.device = torch.device(torch._C._get_default_device()), # torch.device('cpu'),
running editable_wheel
creating /tmp/user/21163/pip-wheel-6kdhr3i2/.tmp-p2xpmenf/vllm.egg-info
writing /tmp/user/21163/pip-wheel-6kdhr3i2/.tmp-p2xpmenf/vllm.egg-info/PKG-INFO
writing dependency_links to /tmp/user/21163/pip-wheel-6kdhr3i2/.tmp-p2xpmenf/vllm.egg-info/dependency_links.txt
writing requirements to /tmp/user/21163/pip-wheel-6kdhr3i2/.tmp-p2xpmenf/vllm.egg-info/requires.txt
writing top-level names to /tmp/user/21163/pip-wheel-6kdhr3i2/.tmp-p2xpmenf/vllm.egg-info/top_level.txt
writing manifest file '/tmp/user/21163/pip-wheel-6kdhr3i2/.tmp-p2xpmenf/vllm.egg-info/SOURCES.txt'
reading manifest file '/tmp/user/21163/pip-wheel-6kdhr3i2/.tmp-p2xpmenf/vllm.egg-info/SOURCES.txt'
reading manifest template 'MANIFEST.in'
adding license file 'LICENSE'
writing manifest file '/tmp/user/21163/pip-wheel-6kdhr3i2/.tmp-p2xpmenf/vllm.egg-info/SOURCES.txt'
creating '/tmp/user/21163/pip-wheel-6kdhr3i2/.tmp-p2xpmenf/vllm-0.3.3+cu122.dist-info'
creating /tmp/user/21163/pip-wheel-6kdhr3i2/.tmp-p2xpmenf/vllm-0.3.3+cu122.dist-info/WHEEL
running build_py
running build_ext
-- The CXX compiler identification is GNU 9.4.0
-- Detecting CXX compiler ABI info
-- Detecting CXX compiler ABI info - done
-- Check for working CXX compiler: /usr/bin/c++ - skipped
-- Detecting CXX compile features
-- Detecting CXX compile features - done
-- Build type: RelWithDebInfo
-- Found Python: /lfs/ampere8/0/rschaef/miniconda3/envs/model_collapse_env/bin/python3.10 (found version "3.10.14") found components: Interpreter Development.Module
-- Found python matching: /lfs/ampere8/0/rschaef/miniconda3/envs/model_collapse_env/bin/python3.10.
-- Found CUDA: /usr/local/cuda (found version "12.2")
-- The CUDA compiler identification is unknown
CMake Error at /tmp/user/21163/pip-build-env-ygusfc74/overlay/lib/python3.10/site-packages/cmake/data/share/cmake-3.28/Modules/CMakeDetermineCUDACompiler.cmake:270 (message):
Failed to detect a default CUDA architecture.
Compiler output:
Call Stack (most recent call first):
/tmp/user/21163/pip-build-env-ygusfc74/overlay/lib/python3.10/site-packages/torch/share/cmake/Caffe2/public/cuda.cmake:47 (enable_language)
/tmp/user/21163/pip-build-env-ygusfc74/overlay/lib/python3.10/site-packages/torch/share/cmake/Caffe2/Caffe2Config.cmake:87 (include)
/tmp/user/21163/pip-build-env-ygusfc74/overlay/lib/python3.10/site-packages/torch/share/cmake/Torch/TorchConfig.cmake:68 (find_package)
CMakeLists.txt:64 (find_package)
-- Configuring incomplete, errors occurred!
Traceback (most recent call last):
File "/tmp/user/21163/pip-build-env-ygusfc74/overlay/lib/python3.10/site-packages/setuptools/command/editable_wheel.py", line 155, in run
self._create_wheel_file(bdist_wheel)
File "/tmp/user/21163/pip-build-env-ygusfc74/overlay/lib/python3.10/site-packages/setuptools/command/editable_wheel.py", line 357, in _create_wheel_file
files, mapping = self._run_build_commands(dist_name, unpacked, lib, tmp)
File "/tmp/user/21163/pip-build-env-ygusfc74/overlay/lib/python3.10/site-packages/setuptools/command/editable_wheel.py", line 280, in _run_build_commands
self._run_build_subcommands()
File "/tmp/user/21163/pip-build-env-ygusfc74/overlay/lib/python3.10/site-packages/setuptools/command/editable_wheel.py", line 307, in _run_build_subcommands
self.run_command(name)
File "/tmp/user/21163/pip-build-env-ygusfc74/overlay/lib/python3.10/site-packages/setuptools/_distutils/cmd.py", line 318, in run_command
self.distribution.run_command(command)
File "/tmp/user/21163/pip-build-env-ygusfc74/overlay/lib/python3.10/site-packages/setuptools/dist.py", line 967, in run_command
super().run_command(command)
File "/tmp/user/21163/pip-build-env-ygusfc74/overlay/lib/python3.10/site-packages/setuptools/_distutils/dist.py", line 988, in run_command
cmd_obj.run()
File "/tmp/user/21163/pip-build-env-ygusfc74/overlay/lib/python3.10/site-packages/setuptools/command/build_ext.py", line 91, in run
_build_ext.run(self)
File "/tmp/user/21163/pip-build-env-ygusfc74/overlay/lib/python3.10/site-packages/setuptools/_distutils/command/build_ext.py", line 345, in run
self.build_extensions()
File "<string>", line 164, in build_extensions
File "<string>", line 147, in configure
File "/lfs/ampere8/0/rschaef/miniconda3/envs/model_collapse_env/lib/python3.10/subprocess.py", line 369, in check_call
raise CalledProcessError(retcode, cmd)
subprocess.CalledProcessError: Command '['cmake', '/lfs/ampere8/0/rschaef/KoyejoLab-Revisiting-Model-Collapse/submodules/vllm', '-G', 'Ninja', '-DCMAKE_BUILD_TYPE=RelWithDebInfo', '-DCMAKE_LIBRARY_OUTPUT_DIRECTORY=/tmp/user/21163/tmp66kpgsiv.build-lib/vllm', '-DCMAKE_ARCHIVE_OUTPUT_DIRECTORY=/tmp/user/21163/tmpzvgnump_.build-temp', '-DVLLM_PYTHON_EXECUTABLE=/lfs/ampere8/0/rschaef/miniconda3/envs/model_collapse_env/bin/python3.10', '-DNVCC_THREADS=8', '-DCMAKE_JOB_POOL_COMPILE:STRING=compile', '-DCMAKE_JOB_POOLS:STRING=compile=64']' returned non-zero exit status 1.
/tmp/user/21163/pip-build-env-ygusfc74/overlay/lib/python3.10/site-packages/setuptools/_distutils/dist.py:988: _DebuggingTips: Problem in editable installation.
!!
********************************************************************************
An error happened while installing `vllm` in editable mode.
The following steps are recommended to help debug this problem:
- Try to install the project normally, without using the editable mode.
Does the error still persist?
(If it does, try fixing the problem before attempting the editable mode).
- If you are using binary extensions, make sure you have all OS-level
dependencies installed (e.g. compilers, toolchains, binary libraries, ...).
- Try the latest version of setuptools (maybe the error was already fixed).
- If you (or your project dependencies) are using any setuptools extension
or customization, make sure they support the editable mode.
After following the steps above, if the problem still persists and
you think this is related to how setuptools handles editable installations,
please submit a reproducible example
(see https://stackoverflow.com/help/minimal-reproducible-example) to:
https://github.com/pypa/setuptools/issues
See https://setuptools.pypa.io/en/latest/userguide/development_mode.html for details.
********************************************************************************
!!
cmd_obj.run()
Traceback (most recent call last):
File "/lfs/ampere8/0/rschaef/miniconda3/envs/model_collapse_env/lib/python3.10/site-packages/pip/_vendor/pyproject_hooks/_in_process/_in_process.py", line 353, in <module>
main()
File "/lfs/ampere8/0/rschaef/miniconda3/envs/model_collapse_env/lib/python3.10/site-packages/pip/_vendor/pyproject_hooks/_in_process/_in_process.py", line 335, in main
json_out['return_val'] = hook(**hook_input['kwargs'])
File "/lfs/ampere8/0/rschaef/miniconda3/envs/model_collapse_env/lib/python3.10/site-packages/pip/_vendor/pyproject_hooks/_in_process/_in_process.py", line 273, in build_editable
return hook(wheel_directory, config_settings, metadata_directory)
File "/tmp/user/21163/pip-build-env-ygusfc74/overlay/lib/python3.10/site-packages/setuptools/build_meta.py", line 443, in build_editable
return self._build_with_temp_dir(
File "/tmp/user/21163/pip-build-env-ygusfc74/overlay/lib/python3.10/site-packages/setuptools/build_meta.py", line 395, in _build_with_temp_dir
self.run_setup()
File "/tmp/user/21163/pip-build-env-ygusfc74/overlay/lib/python3.10/site-packages/setuptools/build_meta.py", line 311, in run_setup
exec(code, locals())
File "<string>", line 345, in <module>
File "/tmp/user/21163/pip-build-env-ygusfc74/overlay/lib/python3.10/site-packages/setuptools/__init__.py", line 104, in setup
return distutils.core.setup(**attrs)
File "/tmp/user/21163/pip-build-env-ygusfc74/overlay/lib/python3.10/site-packages/setuptools/_distutils/core.py", line 185, in setup
return run_commands(dist)
File "/tmp/user/21163/pip-build-env-ygusfc74/overlay/lib/python3.10/site-packages/setuptools/_distutils/core.py", line 201, in run_commands
dist.run_commands()
File "/tmp/user/21163/pip-build-env-ygusfc74/overlay/lib/python3.10/site-packages/setuptools/_distutils/dist.py", line 969, in run_commands
self.run_command(cmd)
File "/tmp/user/21163/pip-build-env-ygusfc74/overlay/lib/python3.10/site-packages/setuptools/dist.py", line 967, in run_command
super().run_command(command)
File "/tmp/user/21163/pip-build-env-ygusfc74/overlay/lib/python3.10/site-packages/setuptools/_distutils/dist.py", line 988, in run_command
cmd_obj.run()
File "/tmp/user/21163/pip-build-env-ygusfc74/overlay/lib/python3.10/site-packages/setuptools/command/editable_wheel.py", line 155, in run
self._create_wheel_file(bdist_wheel)
File "/tmp/user/21163/pip-build-env-ygusfc74/overlay/lib/python3.10/site-packages/setuptools/command/editable_wheel.py", line 357, in _create_wheel_file
files, mapping = self._run_build_commands(dist_name, unpacked, lib, tmp)
File "/tmp/user/21163/pip-build-env-ygusfc74/overlay/lib/python3.10/site-packages/setuptools/command/editable_wheel.py", line 280, in _run_build_commands
self._run_build_subcommands()
File "/tmp/user/21163/pip-build-env-ygusfc74/overlay/lib/python3.10/site-packages/setuptools/command/editable_wheel.py", line 307, in _run_build_subcommands
self.run_command(name)
File "/tmp/user/21163/pip-build-env-ygusfc74/overlay/lib/python3.10/site-packages/setuptools/_distutils/cmd.py", line 318, in run_command
self.distribution.run_command(command)
File "/tmp/user/21163/pip-build-env-ygusfc74/overlay/lib/python3.10/site-packages/setuptools/dist.py", line 967, in run_command
super().run_command(command)
File "/tmp/user/21163/pip-build-env-ygusfc74/overlay/lib/python3.10/site-packages/setuptools/_distutils/dist.py", line 988, in run_command
cmd_obj.run()
File "/tmp/user/21163/pip-build-env-ygusfc74/overlay/lib/python3.10/site-packages/setuptools/command/build_ext.py", line 91, in run
_build_ext.run(self)
File "/tmp/user/21163/pip-build-env-ygusfc74/overlay/lib/python3.10/site-packages/setuptools/_distutils/command/build_ext.py", line 345, in run
self.build_extensions()
File "<string>", line 164, in build_extensions
File "<string>", line 147, in configure
File "/lfs/ampere8/0/rschaef/miniconda3/envs/model_collapse_env/lib/python3.10/subprocess.py", line 369, in check_call
raise CalledProcessError(retcode, cmd)
subprocess.CalledProcessError: Command '['cmake', '/lfs/ampere8/0/rschaef/KoyejoLab-Revisiting-Model-Collapse/submodules/vllm', '-G', 'Ninja', '-DCMAKE_BUILD_TYPE=RelWithDebInfo', '-DCMAKE_LIBRARY_OUTPUT_DIRECTORY=/tmp/user/21163/tmp66kpgsiv.build-lib/vllm', '-DCMAKE_ARCHIVE_OUTPUT_DIRECTORY=/tmp/user/21163/tmpzvgnump_.build-temp', '-DVLLM_PYTHON_EXECUTABLE=/lfs/ampere8/0/rschaef/miniconda3/envs/model_collapse_env/bin/python3.10', '-DNVCC_THREADS=8', '-DCMAKE_JOB_POOL_COMPILE:STRING=compile', '-DCMAKE_JOB_POOLS:STRING=compile=64']' returned non-zero exit status 1.
[end of output]
note: This error originates from a subprocess, and is likely not a problem with pip.
ERROR: Failed building editable for vllm
Failed to build vllm
ERROR: Could not build wheels for vllm, which is required to install pyproject.toml-based projects
Failed to detect a default CUDA architecture.
You don't have a valid cuda installation. Try which nvcc
, and you can see it will fail.
You can install one by conda install cuda=12.1 -c nvidia -y
.
@youkaichao thanks for help! New error:
Building wheels for collected packages: vllm
Building editable for vllm (pyproject.toml) ... error
error: subprocess-exited-with-error
× Building editable for vllm (pyproject.toml) did not run successfully.
│ exit code: 1
╰─> [175 lines of output]
/tmp/user/21163/pip-build-env-4trpsjo5/overlay/lib/python3.10/site-packages/torch/nn/modules/transformer.py:20: UserWarning: Failed to initialize NumPy: No module named 'numpy' (Triggered internally at ../torch/csrc/utils/tensor_numpy.cpp:84.)
device: torch.device = torch.device(torch._C._get_default_device()), # torch.device('cpu'),
running editable_wheel
creating /tmp/user/21163/pip-wheel-_q87sc5u/.tmp-g7p5fhai/vllm.egg-info
writing /tmp/user/21163/pip-wheel-_q87sc5u/.tmp-g7p5fhai/vllm.egg-info/PKG-INFO
writing dependency_links to /tmp/user/21163/pip-wheel-_q87sc5u/.tmp-g7p5fhai/vllm.egg-info/dependency_links.txt
writing requirements to /tmp/user/21163/pip-wheel-_q87sc5u/.tmp-g7p5fhai/vllm.egg-info/requires.txt
writing top-level names to /tmp/user/21163/pip-wheel-_q87sc5u/.tmp-g7p5fhai/vllm.egg-info/top_level.txt
writing manifest file '/tmp/user/21163/pip-wheel-_q87sc5u/.tmp-g7p5fhai/vllm.egg-info/SOURCES.txt'
reading manifest file '/tmp/user/21163/pip-wheel-_q87sc5u/.tmp-g7p5fhai/vllm.egg-info/SOURCES.txt'
reading manifest template 'MANIFEST.in'
adding license file 'LICENSE'
writing manifest file '/tmp/user/21163/pip-wheel-_q87sc5u/.tmp-g7p5fhai/vllm.egg-info/SOURCES.txt'
creating '/tmp/user/21163/pip-wheel-_q87sc5u/.tmp-g7p5fhai/vllm-0.3.3+cu124.dist-info'
creating /tmp/user/21163/pip-wheel-_q87sc5u/.tmp-g7p5fhai/vllm-0.3.3+cu124.dist-info/WHEEL
running build_py
running build_ext
-- The CXX compiler identification is GNU 9.4.0
-- Detecting CXX compiler ABI info
-- Detecting CXX compiler ABI info - done
-- Check for working CXX compiler: /usr/bin/c++ - skipped
-- Detecting CXX compile features
-- Detecting CXX compile features - done
-- Build type: RelWithDebInfo
-- Found Python: /lfs/ampere8/0/rschaef/miniconda3/envs/model_collapse_env/bin/python3.10 (found version "3.10.14") found components: Interpreter Development.Module
-- Found python matching: /lfs/ampere8/0/rschaef/miniconda3/envs/model_collapse_env/bin/python3.10.
-- Found CUDA: /lfs/ampere8/0/rschaef/miniconda3/envs/model_collapse_env (found version "12.4")
-- The CUDA compiler identification is NVIDIA 12.4.99
-- Detecting CUDA compiler ABI info
-- Detecting CUDA compiler ABI info - done
-- Check for working CUDA compiler: /lfs/ampere8/0/rschaef/miniconda3/envs/model_collapse_env/bin/nvcc - skipped
-- Detecting CUDA compile features
-- Detecting CUDA compile features - done
-- Found CUDAToolkit: /lfs/ampere8/0/rschaef/miniconda3/envs/model_collapse_env/include (found version "12.4.99")
-- Performing Test CMAKE_HAVE_LIBC_PTHREAD
-- Performing Test CMAKE_HAVE_LIBC_PTHREAD - Failed
-- Looking for pthread_create in pthreads
-- Looking for pthread_create in pthreads - not found
-- Looking for pthread_create in pthread
-- Looking for pthread_create in pthread - found
-- Found Threads: TRUE
-- Caffe2: CUDA detected: 12.4
-- Caffe2: CUDA nvcc is: /lfs/ampere8/0/rschaef/miniconda3/envs/model_collapse_env/bin/nvcc
-- Caffe2: CUDA toolkit directory: /lfs/ampere8/0/rschaef/miniconda3/envs/model_collapse_env
-- Caffe2: Header version is: 12.4
-- /lfs/ampere8/0/rschaef/miniconda3/envs/model_collapse_env/lib/libnvrtc.so shorthash is a438b617
-- USE_CUDNN is set to 0. Compiling without cuDNN support
-- USE_CUSPARSELT is set to 0. Compiling without cuSPARSELt support
-- Autodetected CUDA architecture(s): 8.0 8.0 8.0 8.0 8.0 8.0 8.0 8.0
-- Added CUDA NVCC flags for: -gencode;arch=compute_80,code=sm_80
CMake Warning at /tmp/user/21163/pip-build-env-4trpsjo5/overlay/lib/python3.10/site-packages/torch/share/cmake/Torch/TorchConfig.cmake:22 (message):
static library kineto_LIBRARY-NOTFOUND not found.
Call Stack (most recent call first):
/tmp/user/21163/pip-build-env-4trpsjo5/overlay/lib/python3.10/site-packages/torch/share/cmake/Torch/TorchConfig.cmake:127 (append_torchlib_if_found)
CMakeLists.txt:64 (find_package)
-- Found Torch: /tmp/user/21163/pip-build-env-4trpsjo5/overlay/lib/python3.10/site-packages/torch/lib/libtorch.so
-- CUDA supported arches: 7.0;7.5;8.0;8.6;8.9;9.0
-- CUDA target arches: 80
-- Punica target arches: 80
-- Enabling C extension.
-- Enabling moe extension.
-- Configuring done (6.6s)
CMake Error: The following variables are used in this project, but they are set to NOTFOUND.
Please set them or make sure they are set and tested correctly in the CMake files:
LIBNVTOOLSEXT
linked by target "_punica_C" in directory /lfs/ampere8/0/rschaef/KoyejoLab-Revisiting-Model-Collapse/submodules/vllm
linked by target "_moe_C" in directory /lfs/ampere8/0/rschaef/KoyejoLab-Revisiting-Model-Collapse/submodules/vllm
linked by target "_C" in directory /lfs/ampere8/0/rschaef/KoyejoLab-Revisiting-Model-Collapse/submodules/vllm
-- Generating done (0.0s)
CMake Generate step failed. Build files cannot be regenerated correctly.
Traceback (most recent call last):
File "/tmp/user/21163/pip-build-env-4trpsjo5/overlay/lib/python3.10/site-packages/setuptools/command/editable_wheel.py", line 155, in run
self._create_wheel_file(bdist_wheel)
File "/tmp/user/21163/pip-build-env-4trpsjo5/overlay/lib/python3.10/site-packages/setuptools/command/editable_wheel.py", line 357, in _create_wheel_file
files, mapping = self._run_build_commands(dist_name, unpacked, lib, tmp)
File "/tmp/user/21163/pip-build-env-4trpsjo5/overlay/lib/python3.10/site-packages/setuptools/command/editable_wheel.py", line 280, in _run_build_commands
self._run_build_subcommands()
File "/tmp/user/21163/pip-build-env-4trpsjo5/overlay/lib/python3.10/site-packages/setuptools/command/editable_wheel.py", line 307, in _run_build_subcommands
self.run_command(name)
File "/tmp/user/21163/pip-build-env-4trpsjo5/overlay/lib/python3.10/site-packages/setuptools/_distutils/cmd.py", line 318, in run_command
self.distribution.run_command(command)
File "/tmp/user/21163/pip-build-env-4trpsjo5/overlay/lib/python3.10/site-packages/setuptools/dist.py", line 967, in run_command
super().run_command(command)
File "/tmp/user/21163/pip-build-env-4trpsjo5/overlay/lib/python3.10/site-packages/setuptools/_distutils/dist.py", line 988, in run_command
cmd_obj.run()
File "/tmp/user/21163/pip-build-env-4trpsjo5/overlay/lib/python3.10/site-packages/setuptools/command/build_ext.py", line 91, in run
_build_ext.run(self)
File "/tmp/user/21163/pip-build-env-4trpsjo5/overlay/lib/python3.10/site-packages/setuptools/_distutils/command/build_ext.py", line 345, in run
self.build_extensions()
File "<string>", line 164, in build_extensions
File "<string>", line 147, in configure
File "/lfs/ampere8/0/rschaef/miniconda3/envs/model_collapse_env/lib/python3.10/subprocess.py", line 369, in check_call
raise CalledProcessError(retcode, cmd)
subprocess.CalledProcessError: Command '['cmake', '/lfs/ampere8/0/rschaef/KoyejoLab-Revisiting-Model-Collapse/submodules/vllm', '-G', 'Ninja', '-DCMAKE_BUILD_TYPE=RelWithDebInfo', '-DCMAKE_LIBRARY_OUTPUT_DIRECTORY=/tmp/user/21163/tmppm34pm7i.build-lib/vllm', '-DCMAKE_ARCHIVE_OUTPUT_DIRECTORY=/tmp/user/21163/tmpt2b60pm3.build-temp', '-DVLLM_PYTHON_EXECUTABLE=/lfs/ampere8/0/rschaef/miniconda3/envs/model_collapse_env/bin/python3.10', '-DNVCC_THREADS=8', '-DCMAKE_JOB_POOL_COMPILE:STRING=compile', '-DCMAKE_JOB_POOLS:STRING=compile=64']' returned non-zero exit status 1.
/tmp/user/21163/pip-build-env-4trpsjo5/overlay/lib/python3.10/site-packages/setuptools/_distutils/dist.py:988: _DebuggingTips: Problem in editable installation.
!!
I'm not sure why it says 12.4. I'm uninstalling and trying conda install nvidia/label/cuda-12.2.0::cuda
instead.
$ nvcc --version
nvcc: NVIDIA (R) Cuda compiler driver
Copyright (c) 2005-2023 NVIDIA Corporation
Built on Tue_Jun_13_19:16:58_PDT_2023
Cuda compilation tools, release 12.2, V12.2.91
Build cuda_12.2.r12.2/compiler.32965470_0
This matches nvidia-smi
:
| NVIDIA-SMI 535.54.03 Driver Version: 535.54.03 CUDA Version: 12.2 |
Now trying pip install -e .
again.
Failed again:
Building wheels for collected packages: vllm
Building editable for vllm (pyproject.toml) ... error
error: subprocess-exited-with-error
× Building editable for vllm (pyproject.toml) did not run successfully.
│ exit code: 1
╰─> [175 lines of output]
/tmp/user/21163/pip-build-env-rnwillx9/overlay/lib/python3.10/site-packages/torch/nn/modules/transformer.py:20: UserWarning: Failed to initialize NumPy: No module named 'numpy' (Triggered internally at ../torch/csrc/utils/tensor_numpy.cpp:84.)
device: torch.device = torch.device(torch._C._get_default_device()), # torch.device('cpu'),
running editable_wheel
creating /tmp/user/21163/pip-wheel-89cl6la1/.tmp-r3mm9nv7/vllm.egg-info
writing /tmp/user/21163/pip-wheel-89cl6la1/.tmp-r3mm9nv7/vllm.egg-info/PKG-INFO
writing dependency_links to /tmp/user/21163/pip-wheel-89cl6la1/.tmp-r3mm9nv7/vllm.egg-info/dependency_links.txt
writing requirements to /tmp/user/21163/pip-wheel-89cl6la1/.tmp-r3mm9nv7/vllm.egg-info/requires.txt
writing top-level names to /tmp/user/21163/pip-wheel-89cl6la1/.tmp-r3mm9nv7/vllm.egg-info/top_level.txt
writing manifest file '/tmp/user/21163/pip-wheel-89cl6la1/.tmp-r3mm9nv7/vllm.egg-info/SOURCES.txt'
reading manifest file '/tmp/user/21163/pip-wheel-89cl6la1/.tmp-r3mm9nv7/vllm.egg-info/SOURCES.txt'
reading manifest template 'MANIFEST.in'
adding license file 'LICENSE'
writing manifest file '/tmp/user/21163/pip-wheel-89cl6la1/.tmp-r3mm9nv7/vllm.egg-info/SOURCES.txt'
creating '/tmp/user/21163/pip-wheel-89cl6la1/.tmp-r3mm9nv7/vllm-0.3.3+cu122.dist-info'
creating /tmp/user/21163/pip-wheel-89cl6la1/.tmp-r3mm9nv7/vllm-0.3.3+cu122.dist-info/WHEEL
running build_py
running build_ext
-- The CXX compiler identification is GNU 9.4.0
-- Detecting CXX compiler ABI info
-- Detecting CXX compiler ABI info - done
-- Check for working CXX compiler: /usr/bin/c++ - skipped
-- Detecting CXX compile features
-- Detecting CXX compile features - done
-- Build type: RelWithDebInfo
-- Found Python: /lfs/ampere8/0/rschaef/miniconda3/envs/model_collapse_env/bin/python3.10 (found version "3.10.14") found components: Interpreter Development.Module
-- Found python matching: /lfs/ampere8/0/rschaef/miniconda3/envs/model_collapse_env/bin/python3.10.
-- Found CUDA: /lfs/ampere8/0/rschaef/miniconda3/envs/model_collapse_env (found version "12.2")
-- The CUDA compiler identification is NVIDIA 12.2.91
-- Detecting CUDA compiler ABI info
-- Detecting CUDA compiler ABI info - done
-- Check for working CUDA compiler: /lfs/ampere8/0/rschaef/miniconda3/envs/model_collapse_env/bin/nvcc - skipped
-- Detecting CUDA compile features
-- Detecting CUDA compile features - done
-- Found CUDAToolkit: /lfs/ampere8/0/rschaef/miniconda3/envs/model_collapse_env/include (found version "12.2.91")
-- Performing Test CMAKE_HAVE_LIBC_PTHREAD
-- Performing Test CMAKE_HAVE_LIBC_PTHREAD - Failed
-- Looking for pthread_create in pthreads
-- Looking for pthread_create in pthreads - not found
-- Looking for pthread_create in pthread
-- Looking for pthread_create in pthread - found
-- Found Threads: TRUE
-- Caffe2: CUDA detected: 12.2
-- Caffe2: CUDA nvcc is: /lfs/ampere8/0/rschaef/miniconda3/envs/model_collapse_env/bin/nvcc
-- Caffe2: CUDA toolkit directory: /lfs/ampere8/0/rschaef/miniconda3/envs/model_collapse_env
-- Caffe2: Header version is: 12.2
-- /lfs/ampere8/0/rschaef/miniconda3/envs/model_collapse_env/lib/libnvrtc.so shorthash is 31d24739
-- USE_CUDNN is set to 0. Compiling without cuDNN support
-- USE_CUSPARSELT is set to 0. Compiling without cuSPARSELt support
-- Autodetected CUDA architecture(s): 8.0 8.0 8.0 8.0 8.0 8.0 8.0 8.0
-- Added CUDA NVCC flags for: -gencode;arch=compute_80,code=sm_80
CMake Warning at /tmp/user/21163/pip-build-env-rnwillx9/overlay/lib/python3.10/site-packages/torch/share/cmake/Torch/TorchConfig.cmake:22 (message):
static library kineto_LIBRARY-NOTFOUND not found.
Call Stack (most recent call first):
/tmp/user/21163/pip-build-env-rnwillx9/overlay/lib/python3.10/site-packages/torch/share/cmake/Torch/TorchConfig.cmake:127 (append_torchlib_if_found)
CMakeLists.txt:64 (find_package)
-- Found Torch: /tmp/user/21163/pip-build-env-rnwillx9/overlay/lib/python3.10/site-packages/torch/lib/libtorch.so
-- CUDA supported arches: 7.0;7.5;8.0;8.6;8.9;9.0
-- CUDA target arches: 80
-- Punica target arches: 80
-- Enabling C extension.
-- Enabling moe extension.
-- Configuring done (6.7s)
CMake Error: The following variables are used in this project, but they are set to NOTFOUND.
Please set them or make sure they are set and tested correctly in the CMake files:
LIBNVTOOLSEXT
linked by target "_punica_C" in directory /lfs/ampere8/0/rschaef/KoyejoLab-Revisiting-Model-Collapse/submodules/vllm
linked by target "_moe_C" in directory /lfs/ampere8/0/rschaef/KoyejoLab-Revisiting-Model-Collapse/submodules/vllm
linked by target "_C" in directory /lfs/ampere8/0/rschaef/KoyejoLab-Revisiting-Model-Collapse/submodules/vllm
-- Generating done (0.0s)
CMake Generate step failed. Build files cannot be regenerated correctly.
Traceback (most recent call last):
File "/tmp/user/21163/pip-build-env-rnwillx9/overlay/lib/python3.10/site-packages/setuptools/command/editable_wheel.py", line 155, in run
self._create_wheel_file(bdist_wheel)
File "/tmp/user/21163/pip-build-env-rnwillx9/overlay/lib/python3.10/site-packages/setuptools/command/editable_wheel.py", line 357, in _create_wheel_file
files, mapping = self._run_build_commands(dist_name, unpacked, lib, tmp)
File "/tmp/user/21163/pip-build-env-rnwillx9/overlay/lib/python3.10/site-packages/setuptools/command/editable_wheel.py", line 280, in _run_build_commands
self._run_build_subcommands()
File "/tmp/user/21163/pip-build-env-rnwillx9/overlay/lib/python3.10/site-packages/setuptools/command/editable_wheel.py", line 307, in _run_build_subcommands
self.run_command(name)
File "/tmp/user/21163/pip-build-env-rnwillx9/overlay/lib/python3.10/site-packages/setuptools/_distutils/cmd.py", line 318, in run_command
self.distribution.run_command(command)
File "/tmp/user/21163/pip-build-env-rnwillx9/overlay/lib/python3.10/site-packages/setuptools/dist.py", line 967, in run_command
super().run_command(command)
File "/tmp/user/21163/pip-build-env-rnwillx9/overlay/lib/python3.10/site-packages/setuptools/_distutils/dist.py", line 988, in run_command
cmd_obj.run()
File "/tmp/user/21163/pip-build-env-rnwillx9/overlay/lib/python3.10/site-packages/setuptools/command/build_ext.py", line 91, in run
_build_ext.run(self)
File "/tmp/user/21163/pip-build-env-rnwillx9/overlay/lib/python3.10/site-packages/setuptools/_distutils/command/build_ext.py", line 345, in run
self.build_extensions()
File "<string>", line 164, in build_extensions
File "<string>", line 147, in configure
File "/lfs/ampere8/0/rschaef/miniconda3/envs/model_collapse_env/lib/python3.10/subprocess.py", line 369, in check_call
raise CalledProcessError(retcode, cmd)
subprocess.CalledProcessError: Command '['cmake', '/lfs/ampere8/0/rschaef/KoyejoLab-Revisiting-Model-Collapse/submodules/vllm', '-G', 'Ninja', '-DCMAKE_BUILD_TYPE=RelWithDebInfo', '-DCMAKE_LIBRARY_OUTPUT_DIRECTORY=/tmp/user/21163/tmpi8zwfa0x.build-lib/vllm', '-DCMAKE_ARCHIVE_OUTPUT_DIRECTORY=/tmp/user/21163/tmptu5etr5v.build-temp', '-DVLLM_PYTHON_EXECUTABLE=/lfs/ampere8/0/rschaef/miniconda3/envs/model_collapse_env/bin/python3.10', '-DNVCC_THREADS=8', '-DCMAKE_JOB_POOL_COMPILE:STRING=compile', '-DCMAKE_JOB_POOLS:STRING=compile=64']' returned non-zero exit status 1.
/tmp/user/21163/pip-build-env-rnwillx9/overlay/lib/python3.10/site-packages/setuptools/_distutils/dist.py:988: _DebuggingTips: Problem in editable installation.
!!
********************************************************************************
An error happened while installing `vllm` in editable mode.
The following steps are recommended to help debug this problem:
- Try to install the project normally, without using the editable mode.
Does the error still persist?
(If it does, try fixing the problem before attempting the editable mode).
- If you are using binary extensions, make sure you have all OS-level
dependencies installed (e.g. compilers, toolchains, binary libraries, ...).
- Try the latest version of setuptools (maybe the error was already fixed).
- If you (or your project dependencies) are using any setuptools extension
or customization, make sure they support the editable mode.
After following the steps above, if the problem still persists and
you think this is related to how setuptools handles editable installations,
please submit a reproducible example
(see https://stackoverflow.com/help/minimal-reproducible-example) to:
https://github.com/pypa/setuptools/issues
See https://setuptools.pypa.io/en/latest/userguide/development_mode.html for details.
********************************************************************************
!!
cmd_obj.run()
CMake Warning at /tmp/user/21163/pip-build-env-rnwillx9/overlay/lib/python3.10/site-packages/torch/share/cmake/Torch/TorchConfig.cmake:22 (message): static library kineto_LIBRARY-NOTFOUND not found.
It seems to be a problem of yout pytorch environment. How did you install pytorch?
How did you install pytorch?
Following the instructions on the pytorch website:
conda install pytorch torchvision torchaudio pytorch-cuda=12.1 -c pytorch -c nvidia
I promise I'm not trying to do something weird. I'm literally trying to install the most vanilla versions of everything.
You can try to use our docker image and see if it works for you: docker pull vllm/vllm-openai:v0.3.3
The script to build the image is also available https://github.com/vllm-project/vllm/blob/main/Dockerfile .
Here's what I just tried:
conda remove -n model_collapse_env --all
conda clean --all
rm -rf .cache
conda create -n model_collapse_env python=3.11
conda activate model_collapse_env
conda install pytorch torchvision torchaudio pytorch-cuda=12.1 -c pytorch -c nvidia
nvidia-smi
- looks finevllm
directory, git pull
and install again: pip install --no-cache-dir --force-reinstall -e .
The error:
$ pip install --no-cache-dir --force-reinstall -e .
Obtaining file:///lfs/ampere8/0/rschaef/KoyejoLab-Revisiting-Model-Collapse/submodules/vllm
Installing build dependencies ... done
Checking if build backend supports build_editable ... done
Getting requirements to build editable ... done
Preparing editable metadata (pyproject.toml) ... done
Collecting cmake>=3.21 (from vllm==0.3.3+cu122)
Downloading cmake-3.28.4-py2.py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl.metadata (6.3 kB)
Collecting ninja (from vllm==0.3.3+cu122)
Downloading ninja-1.11.1.1-py2.py3-none-manylinux1_x86_64.manylinux_2_5_x86_64.whl.metadata (5.3 kB)
Collecting psutil (from vllm==0.3.3+cu122)
Downloading psutil-5.9.8-cp36-abi3-manylinux_2_12_x86_64.manylinux2010_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (21 kB)
Collecting ray>=2.9 (from vllm==0.3.3+cu122)
Downloading ray-2.10.0-cp311-cp311-manylinux2014_x86_64.whl.metadata (13 kB)
Collecting sentencepiece (from vllm==0.3.3+cu122)
Downloading sentencepiece-0.2.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (7.7 kB)
Collecting numpy (from vllm==0.3.3+cu122)
Downloading numpy-1.26.4-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (61 kB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 61.0/61.0 kB 9.7 MB/s eta 0:00:00
Collecting torch==2.1.2 (from vllm==0.3.3+cu122)
Downloading torch-2.1.2-cp311-cp311-manylinux1_x86_64.whl.metadata (25 kB)
Collecting transformers>=4.39.1 (from vllm==0.3.3+cu122)
Downloading transformers-4.39.1-py3-none-any.whl.metadata (134 kB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 134.8/134.8 kB 20.5 MB/s eta 0:00:00
Collecting xformers==0.0.23.post1 (from vllm==0.3.3+cu122)
Downloading xformers-0.0.23.post1-cp311-cp311-manylinux2014_x86_64.whl.metadata (1.0 kB)
Collecting fastapi (from vllm==0.3.3+cu122)
Downloading fastapi-0.110.0-py3-none-any.whl.metadata (25 kB)
Collecting uvicorn[standard] (from vllm==0.3.3+cu122)
Downloading uvicorn-0.29.0-py3-none-any.whl.metadata (6.3 kB)
Collecting pydantic>=2.0 (from vllm==0.3.3+cu122)
Downloading pydantic-2.6.4-py3-none-any.whl.metadata (85 kB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 85.1/85.1 kB 130.5 MB/s eta 0:00:00
Collecting prometheus-client>=0.18.0 (from vllm==0.3.3+cu122)
Downloading prometheus_client-0.20.0-py3-none-any.whl.metadata (1.8 kB)
Collecting pynvml==11.5.0 (from vllm==0.3.3+cu122)
Downloading pynvml-11.5.0-py3-none-any.whl.metadata (7.8 kB)
Collecting triton>=2.1.0 (from vllm==0.3.3+cu122)
Downloading triton-2.2.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (1.4 kB)
Collecting outlines==0.0.34 (from vllm==0.3.3+cu122)
Downloading outlines-0.0.34-py3-none-any.whl.metadata (13 kB)
Collecting cupy-cuda12x==12.1.0 (from vllm==0.3.3+cu122)
Downloading cupy_cuda12x-12.1.0-cp311-cp311-manylinux2014_x86_64.whl.metadata (2.6 kB)
Collecting fastrlock>=0.5 (from cupy-cuda12x==12.1.0->vllm==0.3.3+cu122)
Downloading fastrlock-0.8.2-cp311-cp311-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_28_x86_64.whl.metadata (9.3 kB)
Collecting interegular (from outlines==0.0.34->vllm==0.3.3+cu122)
Downloading interegular-0.3.3-py37-none-any.whl.metadata (3.0 kB)
Collecting jinja2 (from outlines==0.0.34->vllm==0.3.3+cu122)
Downloading Jinja2-3.1.3-py3-none-any.whl.metadata (3.3 kB)
Collecting lark (from outlines==0.0.34->vllm==0.3.3+cu122)
Downloading lark-1.1.9-py3-none-any.whl.metadata (1.9 kB)
Collecting nest-asyncio (from outlines==0.0.34->vllm==0.3.3+cu122)
Downloading nest_asyncio-1.6.0-py3-none-any.whl.metadata (2.8 kB)
Collecting cloudpickle (from outlines==0.0.34->vllm==0.3.3+cu122)
Downloading cloudpickle-3.0.0-py3-none-any.whl.metadata (7.0 kB)
Collecting diskcache (from outlines==0.0.34->vllm==0.3.3+cu122)
Downloading diskcache-5.6.3-py3-none-any.whl.metadata (20 kB)
Collecting scipy (from outlines==0.0.34->vllm==0.3.3+cu122)
Downloading scipy-1.12.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (60 kB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 60.4/60.4 kB 154.2 MB/s eta 0:00:00
Collecting numba (from outlines==0.0.34->vllm==0.3.3+cu122)
Downloading numba-0.59.1-cp311-cp311-manylinux2014_x86_64.manylinux_2_17_x86_64.whl.metadata (2.7 kB)
Collecting joblib (from outlines==0.0.34->vllm==0.3.3+cu122)
Downloading joblib-1.3.2-py3-none-any.whl.metadata (5.4 kB)
Collecting referencing (from outlines==0.0.34->vllm==0.3.3+cu122)
Downloading referencing-0.34.0-py3-none-any.whl.metadata (2.8 kB)
Collecting jsonschema (from outlines==0.0.34->vllm==0.3.3+cu122)
Downloading jsonschema-4.21.1-py3-none-any.whl.metadata (7.8 kB)
Collecting requests (from outlines==0.0.34->vllm==0.3.3+cu122)
Downloading requests-2.31.0-py3-none-any.whl.metadata (4.6 kB)
Collecting filelock (from torch==2.1.2->vllm==0.3.3+cu122)
Downloading filelock-3.13.1-py3-none-any.whl.metadata (2.8 kB)
Collecting typing-extensions (from torch==2.1.2->vllm==0.3.3+cu122)
Downloading typing_extensions-4.10.0-py3-none-any.whl.metadata (3.0 kB)
Collecting sympy (from torch==2.1.2->vllm==0.3.3+cu122)
Downloading sympy-1.12-py3-none-any.whl.metadata (12 kB)
Collecting networkx (from torch==2.1.2->vllm==0.3.3+cu122)
Downloading networkx-3.2.1-py3-none-any.whl.metadata (5.2 kB)
Collecting fsspec (from torch==2.1.2->vllm==0.3.3+cu122)
Downloading fsspec-2024.3.1-py3-none-any.whl.metadata (6.8 kB)
Collecting nvidia-cuda-nvrtc-cu12==12.1.105 (from torch==2.1.2->vllm==0.3.3+cu122)
Downloading nvidia_cuda_nvrtc_cu12-12.1.105-py3-none-manylinux1_x86_64.whl.metadata (1.5 kB)
Collecting nvidia-cuda-runtime-cu12==12.1.105 (from torch==2.1.2->vllm==0.3.3+cu122)
Downloading nvidia_cuda_runtime_cu12-12.1.105-py3-none-manylinux1_x86_64.whl.metadata (1.5 kB)
Collecting nvidia-cuda-cupti-cu12==12.1.105 (from torch==2.1.2->vllm==0.3.3+cu122)
Downloading nvidia_cuda_cupti_cu12-12.1.105-py3-none-manylinux1_x86_64.whl.metadata (1.6 kB)
Collecting nvidia-cudnn-cu12==8.9.2.26 (from torch==2.1.2->vllm==0.3.3+cu122)
Downloading nvidia_cudnn_cu12-8.9.2.26-py3-none-manylinux1_x86_64.whl.metadata (1.6 kB)
Collecting nvidia-cublas-cu12==12.1.3.1 (from torch==2.1.2->vllm==0.3.3+cu122)
Downloading nvidia_cublas_cu12-12.1.3.1-py3-none-manylinux1_x86_64.whl.metadata (1.5 kB)
Collecting nvidia-cufft-cu12==11.0.2.54 (from torch==2.1.2->vllm==0.3.3+cu122)
Downloading nvidia_cufft_cu12-11.0.2.54-py3-none-manylinux1_x86_64.whl.metadata (1.5 kB)
Collecting nvidia-curand-cu12==10.3.2.106 (from torch==2.1.2->vllm==0.3.3+cu122)
Downloading nvidia_curand_cu12-10.3.2.106-py3-none-manylinux1_x86_64.whl.metadata (1.5 kB)
Collecting nvidia-cusolver-cu12==11.4.5.107 (from torch==2.1.2->vllm==0.3.3+cu122)
Downloading nvidia_cusolver_cu12-11.4.5.107-py3-none-manylinux1_x86_64.whl.metadata (1.6 kB)
Collecting nvidia-cusparse-cu12==12.1.0.106 (from torch==2.1.2->vllm==0.3.3+cu122)
Downloading nvidia_cusparse_cu12-12.1.0.106-py3-none-manylinux1_x86_64.whl.metadata (1.6 kB)
Collecting nvidia-nccl-cu12==2.18.1 (from torch==2.1.2->vllm==0.3.3+cu122)
Downloading nvidia_nccl_cu12-2.18.1-py3-none-manylinux1_x86_64.whl.metadata (1.8 kB)
Collecting nvidia-nvtx-cu12==12.1.105 (from torch==2.1.2->vllm==0.3.3+cu122)
Downloading nvidia_nvtx_cu12-12.1.105-py3-none-manylinux1_x86_64.whl.metadata (1.7 kB)
Collecting triton>=2.1.0 (from vllm==0.3.3+cu122)
Downloading triton-2.1.0-0-cp311-cp311-manylinux2014_x86_64.manylinux_2_17_x86_64.whl.metadata (1.3 kB)
Collecting nvidia-nvjitlink-cu12 (from nvidia-cusolver-cu12==11.4.5.107->torch==2.1.2->vllm==0.3.3+cu122)
Downloading nvidia_nvjitlink_cu12-12.4.99-py3-none-manylinux2014_x86_64.whl.metadata (1.5 kB)
Collecting annotated-types>=0.4.0 (from pydantic>=2.0->vllm==0.3.3+cu122)
Downloading annotated_types-0.6.0-py3-none-any.whl.metadata (12 kB)
Collecting pydantic-core==2.16.3 (from pydantic>=2.0->vllm==0.3.3+cu122)
Downloading pydantic_core-2.16.3-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (6.5 kB)
Collecting click>=7.0 (from ray>=2.9->vllm==0.3.3+cu122)
Downloading click-8.1.7-py3-none-any.whl.metadata (3.0 kB)
Collecting msgpack<2.0.0,>=1.0.0 (from ray>=2.9->vllm==0.3.3+cu122)
Downloading msgpack-1.0.8-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (9.1 kB)
Collecting packaging (from ray>=2.9->vllm==0.3.3+cu122)
Downloading packaging-24.0-py3-none-any.whl.metadata (3.2 kB)
Collecting protobuf!=3.19.5,>=3.15.3 (from ray>=2.9->vllm==0.3.3+cu122)
Downloading protobuf-5.26.0-cp37-abi3-manylinux2014_x86_64.whl.metadata (592 bytes)
Collecting pyyaml (from ray>=2.9->vllm==0.3.3+cu122)
Downloading PyYAML-6.0.1-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (2.1 kB)
Collecting aiosignal (from ray>=2.9->vllm==0.3.3+cu122)
Downloading aiosignal-1.3.1-py3-none-any.whl.metadata (4.0 kB)
Collecting frozenlist (from ray>=2.9->vllm==0.3.3+cu122)
Downloading frozenlist-1.4.1-cp311-cp311-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (12 kB)
Collecting huggingface-hub<1.0,>=0.19.3 (from transformers>=4.39.1->vllm==0.3.3+cu122)
Downloading huggingface_hub-0.21.4-py3-none-any.whl.metadata (13 kB)
Collecting regex!=2019.12.17 (from transformers>=4.39.1->vllm==0.3.3+cu122)
Downloading regex-2023.12.25-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (40 kB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 40.9/40.9 kB 113.1 MB/s eta 0:00:00
Collecting tokenizers<0.19,>=0.14 (from transformers>=4.39.1->vllm==0.3.3+cu122)
Downloading tokenizers-0.15.2-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (6.7 kB)
Collecting safetensors>=0.4.1 (from transformers>=4.39.1->vllm==0.3.3+cu122)
Downloading safetensors-0.4.2-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (3.8 kB)
Collecting tqdm>=4.27 (from transformers>=4.39.1->vllm==0.3.3+cu122)
Downloading tqdm-4.66.2-py3-none-any.whl.metadata (57 kB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 57.6/57.6 kB 131.0 MB/s eta 0:00:00
Collecting starlette<0.37.0,>=0.36.3 (from fastapi->vllm==0.3.3+cu122)
Downloading starlette-0.36.3-py3-none-any.whl.metadata (5.9 kB)
Collecting h11>=0.8 (from uvicorn[standard]->vllm==0.3.3+cu122)
Downloading h11-0.14.0-py3-none-any.whl.metadata (8.2 kB)
Collecting httptools>=0.5.0 (from uvicorn[standard]->vllm==0.3.3+cu122)
Downloading httptools-0.6.1-cp311-cp311-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (3.6 kB)
Collecting python-dotenv>=0.13 (from uvicorn[standard]->vllm==0.3.3+cu122)
Downloading python_dotenv-1.0.1-py3-none-any.whl.metadata (23 kB)
Collecting uvloop!=0.15.0,!=0.15.1,>=0.14.0 (from uvicorn[standard]->vllm==0.3.3+cu122)
Downloading uvloop-0.19.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (4.9 kB)
Collecting watchfiles>=0.13 (from uvicorn[standard]->vllm==0.3.3+cu122)
Downloading watchfiles-0.21.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (4.9 kB)
Collecting websockets>=10.4 (from uvicorn[standard]->vllm==0.3.3+cu122)
Downloading websockets-12.0-cp311-cp311-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (6.6 kB)
Collecting anyio<5,>=3.4.0 (from starlette<0.37.0,>=0.36.3->fastapi->vllm==0.3.3+cu122)
Downloading anyio-4.3.0-py3-none-any.whl.metadata (4.6 kB)
Collecting MarkupSafe>=2.0 (from jinja2->outlines==0.0.34->vllm==0.3.3+cu122)
Downloading MarkupSafe-2.1.5-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (3.0 kB)
Collecting attrs>=22.2.0 (from jsonschema->outlines==0.0.34->vllm==0.3.3+cu122)
Downloading attrs-23.2.0-py3-none-any.whl.metadata (9.5 kB)
Collecting jsonschema-specifications>=2023.03.6 (from jsonschema->outlines==0.0.34->vllm==0.3.3+cu122)
Downloading jsonschema_specifications-2023.12.1-py3-none-any.whl.metadata (3.0 kB)
Collecting rpds-py>=0.7.1 (from jsonschema->outlines==0.0.34->vllm==0.3.3+cu122)
Downloading rpds_py-0.18.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (4.1 kB)
Collecting llvmlite<0.43,>=0.42.0dev0 (from numba->outlines==0.0.34->vllm==0.3.3+cu122)
Downloading llvmlite-0.42.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (4.8 kB)
Collecting charset-normalizer<4,>=2 (from requests->outlines==0.0.34->vllm==0.3.3+cu122)
Downloading charset_normalizer-3.3.2-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (33 kB)
Collecting idna<4,>=2.5 (from requests->outlines==0.0.34->vllm==0.3.3+cu122)
Downloading idna-3.6-py3-none-any.whl.metadata (9.9 kB)
Collecting urllib3<3,>=1.21.1 (from requests->outlines==0.0.34->vllm==0.3.3+cu122)
Downloading urllib3-2.2.1-py3-none-any.whl.metadata (6.4 kB)
Collecting certifi>=2017.4.17 (from requests->outlines==0.0.34->vllm==0.3.3+cu122)
Downloading certifi-2024.2.2-py3-none-any.whl.metadata (2.2 kB)
Collecting mpmath>=0.19 (from sympy->torch==2.1.2->vllm==0.3.3+cu122)
Downloading mpmath-1.3.0-py3-none-any.whl.metadata (8.6 kB)
Collecting sniffio>=1.1 (from anyio<5,>=3.4.0->starlette<0.37.0,>=0.36.3->fastapi->vllm==0.3.3+cu122)
Downloading sniffio-1.3.1-py3-none-any.whl.metadata (3.9 kB)
Downloading cupy_cuda12x-12.1.0-cp311-cp311-manylinux2014_x86_64.whl (83.5 MB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 83.5/83.5 MB 155.4 MB/s eta 0:00:00
Downloading outlines-0.0.34-py3-none-any.whl (76 kB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 76.5/76.5 kB 145.6 MB/s eta 0:00:00
Downloading pynvml-11.5.0-py3-none-any.whl (53 kB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 53.1/53.1 kB 129.1 MB/s eta 0:00:00
Downloading torch-2.1.2-cp311-cp311-manylinux1_x86_64.whl (670.2 MB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 670.2/670.2 MB 203.2 MB/s eta 0:00:00
Downloading triton-2.1.0-0-cp311-cp311-manylinux2014_x86_64.manylinux_2_17_x86_64.whl (89.2 MB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 89.2/89.2 MB 200.1 MB/s eta 0:00:00
Downloading xformers-0.0.23.post1-cp311-cp311-manylinux2014_x86_64.whl (213.0 MB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 213.0/213.0 MB 166.5 MB/s eta 0:00:00
Downloading nvidia_cublas_cu12-12.1.3.1-py3-none-manylinux1_x86_64.whl (410.6 MB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 410.6/410.6 MB 223.2 MB/s eta 0:00:00
Downloading nvidia_cuda_cupti_cu12-12.1.105-py3-none-manylinux1_x86_64.whl (14.1 MB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 14.1/14.1 MB 226.2 MB/s eta 0:00:00
Downloading nvidia_cuda_nvrtc_cu12-12.1.105-py3-none-manylinux1_x86_64.whl (23.7 MB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 23.7/23.7 MB 211.4 MB/s eta 0:00:00
Downloading nvidia_cuda_runtime_cu12-12.1.105-py3-none-manylinux1_x86_64.whl (823 kB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 823.6/823.6 kB 175.8 MB/s eta 0:00:00
Downloading nvidia_cudnn_cu12-8.9.2.26-py3-none-manylinux1_x86_64.whl (731.7 MB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 731.7/731.7 MB 228.1 MB/s eta 0:00:00
Downloading nvidia_cufft_cu12-11.0.2.54-py3-none-manylinux1_x86_64.whl (121.6 MB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 121.6/121.6 MB 224.1 MB/s eta 0:00:00
Downloading nvidia_curand_cu12-10.3.2.106-py3-none-manylinux1_x86_64.whl (56.5 MB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 56.5/56.5 MB 162.2 MB/s eta 0:00:00
Downloading nvidia_cusolver_cu12-11.4.5.107-py3-none-manylinux1_x86_64.whl (124.2 MB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 124.2/124.2 MB 195.3 MB/s eta 0:00:00
Downloading nvidia_cusparse_cu12-12.1.0.106-py3-none-manylinux1_x86_64.whl (196.0 MB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 196.0/196.0 MB 149.3 MB/s eta 0:00:00
Downloading nvidia_nccl_cu12-2.18.1-py3-none-manylinux1_x86_64.whl (209.8 MB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 209.8/209.8 MB 163.6 MB/s eta 0:00:00
Downloading nvidia_nvtx_cu12-12.1.105-py3-none-manylinux1_x86_64.whl (99 kB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 99.1/99.1 kB 151.0 MB/s eta 0:00:00
Downloading cmake-3.28.4-py2.py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl (26.3 MB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 26.3/26.3 MB 164.6 MB/s eta 0:00:00
Downloading numpy-1.26.4-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (18.3 MB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 18.3/18.3 MB 166.4 MB/s eta 0:00:00
Downloading prometheus_client-0.20.0-py3-none-any.whl (54 kB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 54.5/54.5 kB 176.4 MB/s eta 0:00:00
Downloading pydantic-2.6.4-py3-none-any.whl (394 kB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 394.9/394.9 kB 223.6 MB/s eta 0:00:00
Downloading pydantic_core-2.16.3-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (2.2 MB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 2.2/2.2 MB 179.2 MB/s eta 0:00:00
Downloading ray-2.10.0-cp311-cp311-manylinux2014_x86_64.whl (65.6 MB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 65.6/65.6 MB 168.1 MB/s eta 0:00:00
Downloading transformers-4.39.1-py3-none-any.whl (8.8 MB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 8.8/8.8 MB 175.7 MB/s eta 0:00:00
Downloading fastapi-0.110.0-py3-none-any.whl (92 kB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 92.1/92.1 kB 160.7 MB/s eta 0:00:00
Downloading ninja-1.11.1.1-py2.py3-none-manylinux1_x86_64.manylinux_2_5_x86_64.whl (307 kB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 307.2/307.2 kB 170.5 MB/s eta 0:00:00
Downloading psutil-5.9.8-cp36-abi3-manylinux_2_12_x86_64.manylinux2010_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl (288 kB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 288.2/288.2 kB 162.2 MB/s eta 0:00:00
Downloading sentencepiece-0.2.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (1.3 MB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 1.3/1.3 MB 202.6 MB/s eta 0:00:00
Downloading annotated_types-0.6.0-py3-none-any.whl (12 kB)
Downloading click-8.1.7-py3-none-any.whl (97 kB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 97.9/97.9 kB 147.6 MB/s eta 0:00:00
Downloading fastrlock-0.8.2-cp311-cp311-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_28_x86_64.whl (52 kB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 52.6/52.6 kB 140.4 MB/s eta 0:00:00
Downloading h11-0.14.0-py3-none-any.whl (58 kB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 58.3/58.3 kB 147.4 MB/s eta 0:00:00
Downloading httptools-0.6.1-cp311-cp311-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl (318 kB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 318.5/318.5 kB 169.8 MB/s eta 0:00:00
Downloading huggingface_hub-0.21.4-py3-none-any.whl (346 kB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 346.4/346.4 kB 212.2 MB/s eta 0:00:00
Downloading fsspec-2024.3.1-py3-none-any.whl (171 kB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 172.0/172.0 kB 215.5 MB/s eta 0:00:00
Downloading msgpack-1.0.8-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (409 kB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 409.3/409.3 kB 175.6 MB/s eta 0:00:00
Downloading packaging-24.0-py3-none-any.whl (53 kB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 53.5/53.5 kB 139.6 MB/s eta 0:00:00
Downloading protobuf-5.26.0-cp37-abi3-manylinux2014_x86_64.whl (302 kB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 302.8/302.8 kB 172.7 MB/s eta 0:00:00
Downloading python_dotenv-1.0.1-py3-none-any.whl (19 kB)
Downloading PyYAML-6.0.1-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (757 kB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 757.7/757.7 kB 212.8 MB/s eta 0:00:00
Downloading regex-2023.12.25-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (785 kB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 785.1/785.1 kB 213.4 MB/s eta 0:00:00
Downloading safetensors-0.4.2-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (1.3 MB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 1.3/1.3 MB 213.1 MB/s eta 0:00:00
Downloading starlette-0.36.3-py3-none-any.whl (71 kB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 71.5/71.5 kB 146.4 MB/s eta 0:00:00
Downloading tokenizers-0.15.2-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (3.6 MB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 3.6/3.6 MB 179.9 MB/s eta 0:00:00
Downloading tqdm-4.66.2-py3-none-any.whl (78 kB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 78.3/78.3 kB 150.6 MB/s eta 0:00:00
Downloading typing_extensions-4.10.0-py3-none-any.whl (33 kB)
Downloading uvloop-0.19.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (3.5 MB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 3.5/3.5 MB 180.8 MB/s eta 0:00:00
Downloading watchfiles-0.21.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (1.3 MB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 1.3/1.3 MB 174.5 MB/s eta 0:00:00
Downloading websockets-12.0-cp311-cp311-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl (130 kB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 130.9/130.9 kB 162.0 MB/s eta 0:00:00
Downloading aiosignal-1.3.1-py3-none-any.whl (7.6 kB)
Downloading frozenlist-1.4.1-cp311-cp311-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl (272 kB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 272.3/272.3 kB 161.1 MB/s eta 0:00:00
Downloading cloudpickle-3.0.0-py3-none-any.whl (20 kB)
Downloading diskcache-5.6.3-py3-none-any.whl (45 kB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 45.5/45.5 kB 168.0 MB/s eta 0:00:00
Downloading filelock-3.13.1-py3-none-any.whl (11 kB)
Downloading interegular-0.3.3-py37-none-any.whl (23 kB)
Downloading Jinja2-3.1.3-py3-none-any.whl (133 kB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 133.2/133.2 kB 169.2 MB/s eta 0:00:00
Downloading joblib-1.3.2-py3-none-any.whl (302 kB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 302.2/302.2 kB 164.8 MB/s eta 0:00:00
Downloading jsonschema-4.21.1-py3-none-any.whl (85 kB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 85.5/85.5 kB 189.3 MB/s eta 0:00:00
Downloading referencing-0.34.0-py3-none-any.whl (26 kB)
Downloading lark-1.1.9-py3-none-any.whl (111 kB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 111.7/111.7 kB 152.4 MB/s eta 0:00:00
Downloading nest_asyncio-1.6.0-py3-none-any.whl (5.2 kB)
Downloading networkx-3.2.1-py3-none-any.whl (1.6 MB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 1.6/1.6 MB 170.3 MB/s eta 0:00:00
Downloading numba-0.59.1-cp311-cp311-manylinux2014_x86_64.manylinux_2_17_x86_64.whl (3.7 MB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 3.7/3.7 MB 175.3 MB/s eta 0:00:00
Downloading requests-2.31.0-py3-none-any.whl (62 kB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 62.6/62.6 kB 183.6 MB/s eta 0:00:00
Downloading scipy-1.12.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (38.4 MB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 38.4/38.4 MB 169.1 MB/s eta 0:00:00
Downloading sympy-1.12-py3-none-any.whl (5.7 MB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 5.7/5.7 MB 178.8 MB/s eta 0:00:00
Downloading uvicorn-0.29.0-py3-none-any.whl (60 kB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 60.8/60.8 kB 146.7 MB/s eta 0:00:00
Downloading anyio-4.3.0-py3-none-any.whl (85 kB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 85.6/85.6 kB 154.5 MB/s eta 0:00:00
Downloading attrs-23.2.0-py3-none-any.whl (60 kB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 60.8/60.8 kB 147.5 MB/s eta 0:00:00
Downloading certifi-2024.2.2-py3-none-any.whl (163 kB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 163.8/163.8 kB 215.2 MB/s eta 0:00:00
Downloading charset_normalizer-3.3.2-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (140 kB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 140.3/140.3 kB 199.7 MB/s eta 0:00:00
Downloading idna-3.6-py3-none-any.whl (61 kB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 61.6/61.6 kB 181.5 MB/s eta 0:00:00
Downloading jsonschema_specifications-2023.12.1-py3-none-any.whl (18 kB)
Downloading llvmlite-0.42.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (43.8 MB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 43.8/43.8 MB 169.1 MB/s eta 0:00:00
Downloading MarkupSafe-2.1.5-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (28 kB)
Downloading mpmath-1.3.0-py3-none-any.whl (536 kB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 536.2/536.2 kB 221.8 MB/s eta 0:00:00
Downloading rpds_py-0.18.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (1.1 MB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 1.1/1.1 MB 164.9 MB/s eta 0:00:00
Downloading urllib3-2.2.1-py3-none-any.whl (121 kB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 121.1/121.1 kB 158.8 MB/s eta 0:00:00
Downloading nvidia_nvjitlink_cu12-12.4.99-py3-none-manylinux2014_x86_64.whl (21.1 MB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 21.1/21.1 MB 170.4 MB/s eta 0:00:00
Downloading sniffio-1.3.1-py3-none-any.whl (10 kB)
Building wheels for collected packages: vllm
Building editable for vllm (pyproject.toml) ... error
error: subprocess-exited-with-error
× Building editable for vllm (pyproject.toml) did not run successfully.
│ exit code: 1
╰─> [152 lines of output]
/tmp/user/21163/pip-build-env-vvupu6fr/overlay/lib/python3.11/site-packages/torch/nn/modules/transformer.py:20: UserWarning: Failed to initialize NumPy: No module named 'numpy' (Triggered internally at ../torch/csrc/utils/tensor_numpy.cpp:84.)
device: torch.device = torch.device(torch._C._get_default_device()), # torch.device('cpu'),
running editable_wheel
creating /tmp/user/21163/pip-wheel-wim8vklg/.tmp-bp23gjik/vllm.egg-info
writing /tmp/user/21163/pip-wheel-wim8vklg/.tmp-bp23gjik/vllm.egg-info/PKG-INFO
writing dependency_links to /tmp/user/21163/pip-wheel-wim8vklg/.tmp-bp23gjik/vllm.egg-info/dependency_links.txt
writing requirements to /tmp/user/21163/pip-wheel-wim8vklg/.tmp-bp23gjik/vllm.egg-info/requires.txt
writing top-level names to /tmp/user/21163/pip-wheel-wim8vklg/.tmp-bp23gjik/vllm.egg-info/top_level.txt
writing manifest file '/tmp/user/21163/pip-wheel-wim8vklg/.tmp-bp23gjik/vllm.egg-info/SOURCES.txt'
reading manifest file '/tmp/user/21163/pip-wheel-wim8vklg/.tmp-bp23gjik/vllm.egg-info/SOURCES.txt'
reading manifest template 'MANIFEST.in'
adding license file 'LICENSE'
writing manifest file '/tmp/user/21163/pip-wheel-wim8vklg/.tmp-bp23gjik/vllm.egg-info/SOURCES.txt'
creating '/tmp/user/21163/pip-wheel-wim8vklg/.tmp-bp23gjik/vllm-0.3.3+cu122.dist-info'
creating /tmp/user/21163/pip-wheel-wim8vklg/.tmp-bp23gjik/vllm-0.3.3+cu122.dist-info/WHEEL
running build_py
running build_ext
-- The CXX compiler identification is GNU 9.4.0
-- Detecting CXX compiler ABI info
-- Detecting CXX compiler ABI info - done
-- Check for working CXX compiler: /usr/bin/c++ - skipped
-- Detecting CXX compile features
-- Detecting CXX compile features - done
-- Build type: RelWithDebInfo
-- Found Python: /lfs/ampere8/0/rschaef/miniconda3/envs/model_collapse_env/bin/python (found version "3.11.8") found components: Interpreter Development.Module
-- Found python matching: /lfs/ampere8/0/rschaef/miniconda3/envs/model_collapse_env/bin/python.
-- Found CUDA: /usr/local/cuda (found version "12.2")
-- The CUDA compiler identification is unknown
CMake Error at /tmp/user/21163/pip-build-env-vvupu6fr/overlay/lib/python3.11/site-packages/cmake/data/share/cmake-3.28/Modules/CMakeDetermineCUDACompiler.cmake:270 (message):
Failed to detect a default CUDA architecture.
Compiler output:
Call Stack (most recent call first):
/tmp/user/21163/pip-build-env-vvupu6fr/overlay/lib/python3.11/site-packages/torch/share/cmake/Caffe2/public/cuda.cmake:47 (enable_language)
/tmp/user/21163/pip-build-env-vvupu6fr/overlay/lib/python3.11/site-packages/torch/share/cmake/Caffe2/Caffe2Config.cmake:87 (include)
/tmp/user/21163/pip-build-env-vvupu6fr/overlay/lib/python3.11/site-packages/torch/share/cmake/Torch/TorchConfig.cmake:68 (find_package)
CMakeLists.txt:64 (find_package)
-- Configuring incomplete, errors occurred!
Traceback (most recent call last):
File "/tmp/user/21163/pip-build-env-vvupu6fr/overlay/lib/python3.11/site-packages/setuptools/command/editable_wheel.py", line 155, in run
self._create_wheel_file(bdist_wheel)
File "/tmp/user/21163/pip-build-env-vvupu6fr/overlay/lib/python3.11/site-packages/setuptools/command/editable_wheel.py", line 357, in _create_wheel_file
files, mapping = self._run_build_commands(dist_name, unpacked, lib, tmp)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/tmp/user/21163/pip-build-env-vvupu6fr/overlay/lib/python3.11/site-packages/setuptools/command/editable_wheel.py", line 280, in _run_build_commands
self._run_build_subcommands()
File "/tmp/user/21163/pip-build-env-vvupu6fr/overlay/lib/python3.11/site-packages/setuptools/command/editable_wheel.py", line 307, in _run_build_subcommands
self.run_command(name)
File "/tmp/user/21163/pip-build-env-vvupu6fr/overlay/lib/python3.11/site-packages/setuptools/_distutils/cmd.py", line 318, in run_command
self.distribution.run_command(command)
File "/tmp/user/21163/pip-build-env-vvupu6fr/overlay/lib/python3.11/site-packages/setuptools/dist.py", line 967, in run_command
super().run_command(command)
File "/tmp/user/21163/pip-build-env-vvupu6fr/overlay/lib/python3.11/site-packages/setuptools/_distutils/dist.py", line 988, in run_command
cmd_obj.run()
File "/tmp/user/21163/pip-build-env-vvupu6fr/overlay/lib/python3.11/site-packages/setuptools/command/build_ext.py", line 91, in run
_build_ext.run(self)
File "/tmp/user/21163/pip-build-env-vvupu6fr/overlay/lib/python3.11/site-packages/setuptools/_distutils/command/build_ext.py", line 345, in run
self.build_extensions()
File "<string>", line 164, in build_extensions
File "<string>", line 147, in configure
File "/lfs/ampere8/0/rschaef/miniconda3/envs/model_collapse_env/lib/python3.11/subprocess.py", line 413, in check_call
raise CalledProcessError(retcode, cmd)
subprocess.CalledProcessError: Command '['cmake', '/lfs/ampere8/0/rschaef/KoyejoLab-Revisiting-Model-Collapse/submodules/vllm', '-G', 'Ninja', '-DCMAKE_BUILD_TYPE=RelWithDebInfo', '-DCMAKE_LIBRARY_OUTPUT_DIRECTORY=/tmp/user/21163/tmp00n843y3.build-lib/vllm', '-DCMAKE_ARCHIVE_OUTPUT_DIRECTORY=/tmp/user/21163/tmpov687z8o.build-temp', '-DVLLM_PYTHON_EXECUTABLE=/lfs/ampere8/0/rschaef/miniconda3/envs/model_collapse_env/bin/python', '-DNVCC_THREADS=8', '-DCMAKE_JOB_POOL_COMPILE:STRING=compile', '-DCMAKE_JOB_POOLS:STRING=compile=64']' returned non-zero exit status 1.
/tmp/user/21163/pip-build-env-vvupu6fr/overlay/lib/python3.11/site-packages/setuptools/_distutils/dist.py:988: _DebuggingTips: Problem in editable installation.
!!
********************************************************************************
An error happened while installing `vllm` in editable mode.
The following steps are recommended to help debug this problem:
- Try to install the project normally, without using the editable mode.
Does the error still persist?
(If it does, try fixing the problem before attempting the editable mode).
- If you are using binary extensions, make sure you have all OS-level
dependencies installed (e.g. compilers, toolchains, binary libraries, ...).
- Try the latest version of setuptools (maybe the error was already fixed).
- If you (or your project dependencies) are using any setuptools extension
or customization, make sure they support the editable mode.
After following the steps above, if the problem still persists and
you think this is related to how setuptools handles editable installations,
please submit a reproducible example
(see https://stackoverflow.com/help/minimal-reproducible-example) to:
https://github.com/pypa/setuptools/issues
See https://setuptools.pypa.io/en/latest/userguide/development_mode.html for details.
********************************************************************************
!!
I don't understand how installing pytorch-cuda doesn't install cuda, but I am now going to try conda install nvidia/label/cuda-12.1.0::cuda
and try installing vllm
again. I'll report back in a few minutes
Error:
Building wheels for collected packages: vllm
Building editable for vllm (pyproject.toml) ... error
error: subprocess-exited-with-error
× Building editable for vllm (pyproject.toml) did not run successfully.
│ exit code: 1
╰─> [155 lines of output]
/tmp/user/21163/pip-build-env-igskn0fq/overlay/lib/python3.11/site-packages/torch/nn/modules/transformer.py:20: UserWarning: Failed to initialize NumPy: No module named 'numpy' (Triggered internally at ../torch/csrc/utils/tensor_numpy.cpp:84.)
device: torch.device = torch.device(torch._C._get_default_device()), # torch.device('cpu'),
running editable_wheel
creating /tmp/user/21163/pip-wheel-2svgs4gy/.tmp-454i10h6/vllm.egg-info
writing /tmp/user/21163/pip-wheel-2svgs4gy/.tmp-454i10h6/vllm.egg-info/PKG-INFO
writing dependency_links to /tmp/user/21163/pip-wheel-2svgs4gy/.tmp-454i10h6/vllm.egg-info/dependency_links.txt
writing requirements to /tmp/user/21163/pip-wheel-2svgs4gy/.tmp-454i10h6/vllm.egg-info/requires.txt
writing top-level names to /tmp/user/21163/pip-wheel-2svgs4gy/.tmp-454i10h6/vllm.egg-info/top_level.txt
writing manifest file '/tmp/user/21163/pip-wheel-2svgs4gy/.tmp-454i10h6/vllm.egg-info/SOURCES.txt'
reading manifest file '/tmp/user/21163/pip-wheel-2svgs4gy/.tmp-454i10h6/vllm.egg-info/SOURCES.txt'
reading manifest template 'MANIFEST.in'
adding license file 'LICENSE'
writing manifest file '/tmp/user/21163/pip-wheel-2svgs4gy/.tmp-454i10h6/vllm.egg-info/SOURCES.txt'
creating '/tmp/user/21163/pip-wheel-2svgs4gy/.tmp-454i10h6/vllm-0.3.3.dist-info'
creating /tmp/user/21163/pip-wheel-2svgs4gy/.tmp-454i10h6/vllm-0.3.3.dist-info/WHEEL
running build_py
running build_ext
-- The CXX compiler identification is GNU 9.4.0
-- Detecting CXX compiler ABI info
-- Detecting CXX compiler ABI info - done
-- Check for working CXX compiler: /usr/bin/c++ - skipped
-- Detecting CXX compile features
-- Detecting CXX compile features - done
-- Build type: RelWithDebInfo
-- Found Python: /lfs/ampere8/0/rschaef/miniconda3/envs/model_collapse_env/bin/python (found version "3.11.8") found components: Interpreter Development.Module
-- Found python matching: /lfs/ampere8/0/rschaef/miniconda3/envs/model_collapse_env/bin/python.
-- Could NOT find CUDA (missing: CUDA_CUDART_LIBRARY) (found version "12.1")
CMake Warning at /tmp/user/21163/pip-build-env-igskn0fq/overlay/lib/python3.11/site-packages/torch/share/cmake/Caffe2/public/cuda.cmake:31 (message):
Caffe2: CUDA cannot be found. Depending on whether you are building Caffe2
or a Caffe2 dependent library, the next warning / error will give you more
info.
Call Stack (most recent call first):
/tmp/user/21163/pip-build-env-igskn0fq/overlay/lib/python3.11/site-packages/torch/share/cmake/Caffe2/Caffe2Config.cmake:87 (include)
/tmp/user/21163/pip-build-env-igskn0fq/overlay/lib/python3.11/site-packages/torch/share/cmake/Torch/TorchConfig.cmake:68 (find_package)
CMakeLists.txt:64 (find_package)
CMake Error at /tmp/user/21163/pip-build-env-igskn0fq/overlay/lib/python3.11/site-packages/torch/share/cmake/Caffe2/Caffe2Config.cmake:91 (message):
Your installed Caffe2 version uses CUDA but I cannot find the CUDA
libraries. Please set the proper CUDA prefixes and / or install CUDA.
Call Stack (most recent call first):
/tmp/user/21163/pip-build-env-igskn0fq/overlay/lib/python3.11/site-packages/torch/share/cmake/Torch/TorchConfig.cmake:68 (find_package)
CMakeLists.txt:64 (find_package)
For more info, nvcc --version
exists: $ nvcc --version
nvcc: NVIDIA (R) Cuda compiler driver
Copyright (c) 2005-2023 NVIDIA Corporation
Built on Tue_Feb__7_19:32:13_PST_2023
Cuda compilation tools, release 12.1, V12.1.66
Build cuda_12.1.r12.1/compiler.32415258_0
If I do conda list
, I see what appears to be CUDA:
cuda 12.1.0 0 nvidia/label/cuda-12.1.0
cuda-cccl 12.1.55 0 nvidia/label/cuda-12.1.0
cuda-command-line-tools 12.1.0 0 nvidia/label/cuda-12.1.0
cuda-compiler 12.1.0 0 nvidia/label/cuda-12.1.0
cuda-cudart 12.1.105 0 nvidia
cuda-cudart-dev 12.1.55 0 nvidia/label/cuda-12.1.0
cuda-cudart-static 12.1.55 0 nvidia/label/cuda-12.1.0
cuda-cuobjdump 12.1.55 0 nvidia/label/cuda-12.1.0
cuda-cupti 12.1.105 0 nvidia
cuda-cupti-static 12.1.62 0 nvidia/label/cuda-12.1.0
cuda-cuxxfilt 12.1.55 0 nvidia/label/cuda-12.1.0
cuda-demo-suite 12.1.55 0 nvidia/label/cuda-12.1.0
cuda-documentation 12.1.55 0 nvidia/label/cuda-12.1.0
cuda-driver-dev 12.1.55 0 nvidia/label/cuda-12.1.0
cuda-gdb 12.1.55 0 nvidia/label/cuda-12.1.0
cuda-libraries 12.1.0 0 nvidia
cuda-libraries-dev 12.1.0 0 nvidia/label/cuda-12.1.0
cuda-libraries-static 12.1.0 0 nvidia/label/cuda-12.1.0
cuda-nsight 12.1.55 0 nvidia/label/cuda-12.1.0
cuda-nsight-compute 12.1.0 0 nvidia/label/cuda-12.1.0
cuda-nvcc 12.1.66 0 nvidia/label/cuda-12.1.0
cuda-nvdisasm 12.1.55 0 nvidia/label/cuda-12.1.0
cuda-nvml-dev 12.1.55 0 nvidia/label/cuda-12.1.0
cuda-nvprof 12.1.55 0 nvidia/label/cuda-12.1.0
cuda-nvprune 12.1.55 0 nvidia/label/cuda-12.1.0
cuda-nvrtc 12.1.105 0 nvidia
cuda-nvrtc-dev 12.1.55 0 nvidia/label/cuda-12.1.0
cuda-nvrtc-static 12.1.55 0 nvidia/label/cuda-12.1.0
cuda-nvtx 12.1.105 0 nvidia
cuda-nvvp 12.1.55 0 nvidia/label/cuda-12.1.0
cuda-opencl 12.4.99 0 nvidia
cuda-opencl-dev 12.1.56 0 nvidia/label/cuda-12.1.0
cuda-profiler-api 12.1.55 0 nvidia/label/cuda-12.1.0
cuda-runtime 12.1.0 0 nvidia
cuda-sanitizer-api 12.1.55 0 nvidia/label/cuda-12.1.0
cuda-toolkit 12.1.0 0 nvidia/label/cuda-12.1.0
cuda-tools 12.1.0 0 nvidia/label/cuda-12.1.0
cuda-visual-tools 12.1.0 0 nvidia/label/cuda-12.1.0
Your environment is complicated and might be broken. Please try the docker image.
I found a solution!!! I just followed https://docs.vllm.ai/en/latest/getting_started/installation.html
pip install vllm
worked right out of the box!!
For people encountering undefined symbol
problem:
TL,DR; please pip install vllm
in a fresh new python environment.
The long answer: this is usually caused by an incompatible pytorch version. By default pip install vllm
will install pre-built binary wheels, which is compiled against a particular pytorch version. Unfortunately, compiled binary objects are not portable across pytorch versions. Either have a fresh new environment and let pip
choose the correct pytorch version for you, or you can build vllm
from source.
I ran into the same issue (LIBNVTOOLSEXT not set) when installing from source.
I fixed it by installing using the v0.3.3 version of the repo. I think this is a problem with the latest main branch.
when build from source, pytorch 2.2 cuda 12.1 :
from vllm._C import ops
ImportError: /workspace/vllm/vllm/_C.cpython-310-x86_64-linux-gnu.so: undefined symbol: _ZN2at4_ops15to_dtype_layout4callERKNS_6TensorEN3c108optionalINS5_10ScalarTypeEEENS6_INS5_6LayoutEEENS6_INS5_6DeviceEEENS6_IbEEbbNS6_INS5_12MemoryFormatEEE
this is miss the ref : at::_ops::to_dtype_layout::call(at::Tensor const&, c10::optional<c10::ScalarType>, c10::optional<c10::Layout>, c10::optional<c10::Device>, c10::optional<bool>, bool, bool, c10::optional<c10::MemoryFormat>)
so is there somewhere to use this ops?
pytorch 2.2 cuda 12.1
pt 2.2 is not supported yet.
I ran into the same issue (LIBNVTOOLSEXT not set) when installing from source.
I fixed it by installing using the v0.3.3 version of the repo. I think this is a problem with the latest main branch.
I had exactly the same issue (LIBNVTOOLSEXT not set when building from source) and rolling back to v0.3.3 solved it. Most likely related to the switch to cmake build (#2830).
pytorch 2.2 cuda 12.1
pt 2.2 is not supported yet.
when will the pytorch 2.2 will be supported ? or how to fix this error ? @youkaichao
from vllm._C import ops
ImportError: /workspace/vllm/vllm/_C.cpython-310-x86_64-linux-gnu.so: undefined symbol: _ZN2at4_ops15to_dtype_layout4callERKNS_6TensorEN3c108optionalINS5_10ScalarTypeEEENS6_INS5_6LayoutEEENS6_INS5_6DeviceEEENS6_IbEEbbNS6_INS5_12MemoryFormatEEE
For people encountering
undefined symbol
problem:TL,DR; please
pip install vllm
in a fresh new python environment.The long answer: this is usually caused by an incompatible pytorch version. By default
pip install vllm
will install pre-built binary wheels, which is compiled against a particular pytorch version. Unfortunately, compiled binary objects are not portable across pytorch versions. Either have a fresh new environment and letpip
choose the correct pytorch version for you, or you can buildvllm
from source.
@youkaichao can you give us a hint of which pytorch version does work? ranges? the highest? anything?
can we request pytorch 2.2? It's the fastest! https://github.com/vllm-project/vllm/issues/3742
@youkaichao I followed @RylanSchaeffer advice and I still get an eror:
ERROR: pip's dependency resolver does not currently take into account all the packages that are installed. This behaviour is the source of the following dependency conflicts.
outlines 0.0.37 requires pydantic>=2.0, but you have pydantic 1.10.13 which is incompatible.
Successfully installed GitPython-3.1.42 Pygments-2.17.2 SecretStorage-3.3.3 absl-py-2.1.0 accelerate-0.28.0 aiohttp-3.9.3 aioprometheus-23.12.0 appdirs-1.4.4 async-timeout-4.0.3 bitsandbytes-0.43.0 cffi-1.16.0 contourpy-1.2.0 cryptography-42.0.5 cycler-0.12.1 datasets-2.18.0 dill-0.3.8 docker-pycreds-0.4.0 docstring-parser-0.16 docutils-0.20.1 einops-0.7.0 fonttools-4.50.0 fsspec-2024.2.0 gitdb-4.0.11 gold-ai-olmpiad-0.1.0 grpcio-1.62.1 importlib-metadata-7.1.0 jaraco.classes-3.3.1 jaraco.context-4.3.0 jaraco.functools-4.0.0 jeepney-0.8.0 keyring-25.0.0 kiwisolver-1.4.5 lark-parser-0.12.0 markdown-3.6 markdown-it-py-3.0.0 matplotlib-3.8.3 mdurl-0.1.2 more-itertools-10.2.0 multidict-6.0.5 multiprocess-0.70.16 nh3-0.2.17 nvidia-nccl-cu12-2.19.3 orjson-3.10.0 pandas-2.2.1 peft-0.10.0 pillow-10.2.0 pkginfo-1.10.0 plotly-5.20.0 progressbar2-4.4.2 protobuf-4.25.3 pyarrow-15.0.2 pyarrow-hotfix-0.6 pycparser-2.21 pydantic-1.10.13 pyparsing-3.1.2 python-dateutil-2.9.0.post0 python-utils-3.8.2 pytz-2024.1 quantile-python-1.1 readme-renderer-43.0 requests-toolbelt-1.0.0 rfc3986-2.0.0 rich-13.7.1 scikit-learn-1.4.1.post1 sentry-sdk-1.44.0 setproctitle-1.3.3 shtab-1.7.1 six-1.16.0 smmap-5.0.1 tenacity-8.2.3 tensorboard-2.16.2 tensorboard-data-server-0.7.2 threadpoolctl-3.4.0 torch-2.2.2 torchaudio-2.2.2 torchvision-0.17.2 triton-2.2.0 trl-0.8.1 twine-5.0.0 tyro-0.7.3 tzdata-2024.1 vllm-0.2.5 wandb-0.16.5 werkzeug-3.0.1 xformers-0.0.25.post1 xxhash-3.4.1 yarl-1.9.4 zipp-3.18.1
Can you let us know precisely the commands you recommend to run? I started new conda env and it threw the above error anyway https://docs.vllm.ai/en/latest/getting_started/installation.html
error still:
_ZN2at4_ops15to_dtype_layout4callERKNS_6TensorEN3c108optionalINS5_10ScalarTypeEEENS6_INS5_6LayoutEEENS6_INS5_6DeviceEEENS6_IbEEbbNS6_INS5_12MemoryFormatEEE
@youkaichao I followed @RylanSchaeffer advice and I still get an eror:
ERROR: pip's dependency resolver does not currently take into account all the packages that are installed. This behaviour is the source of the following dependency conflicts. outlines 0.0.37 requires pydantic>=2.0, but you have pydantic 1.10.13 which is incompatible. Successfully installed GitPython-3.1.42 Pygments-2.17.2 SecretStorage-3.3.3 absl-py-2.1.0 accelerate-0.28.0 aiohttp-3.9.3 aioprometheus-23.12.0 appdirs-1.4.4 async-timeout-4.0.3 bitsandbytes-0.43.0 cffi-1.16.0 contourpy-1.2.0 cryptography-42.0.5 cycler-0.12.1 datasets-2.18.0 dill-0.3.8 docker-pycreds-0.4.0 docstring-parser-0.16 docutils-0.20.1 einops-0.7.0 fonttools-4.50.0 fsspec-2024.2.0 gitdb-4.0.11 gold-ai-olmpiad-0.1.0 grpcio-1.62.1 importlib-metadata-7.1.0 jaraco.classes-3.3.1 jaraco.context-4.3.0 jaraco.functools-4.0.0 jeepney-0.8.0 keyring-25.0.0 kiwisolver-1.4.5 lark-parser-0.12.0 markdown-3.6 markdown-it-py-3.0.0 matplotlib-3.8.3 mdurl-0.1.2 more-itertools-10.2.0 multidict-6.0.5 multiprocess-0.70.16 nh3-0.2.17 nvidia-nccl-cu12-2.19.3 orjson-3.10.0 pandas-2.2.1 peft-0.10.0 pillow-10.2.0 pkginfo-1.10.0 plotly-5.20.0 progressbar2-4.4.2 protobuf-4.25.3 pyarrow-15.0.2 pyarrow-hotfix-0.6 pycparser-2.21 pydantic-1.10.13 pyparsing-3.1.2 python-dateutil-2.9.0.post0 python-utils-3.8.2 pytz-2024.1 quantile-python-1.1 readme-renderer-43.0 requests-toolbelt-1.0.0 rfc3986-2.0.0 rich-13.7.1 scikit-learn-1.4.1.post1 sentry-sdk-1.44.0 setproctitle-1.3.3 shtab-1.7.1 six-1.16.0 smmap-5.0.1 tenacity-8.2.3 tensorboard-2.16.2 tensorboard-data-server-0.7.2 threadpoolctl-3.4.0 torch-2.2.2 torchaudio-2.2.2 torchvision-0.17.2 triton-2.2.0 trl-0.8.1 twine-5.0.0 tyro-0.7.3 tzdata-2024.1 vllm-0.2.5 wandb-0.16.5 werkzeug-3.0.1 xformers-0.0.25.post1 xxhash-3.4.1 yarl-1.9.4 zipp-3.18.1
Can you let us know precisely the commands you recommend to run? I started new conda env and it threw the above error anyway https://docs.vllm.ai/en/latest/getting_started/installation.html
error still:
_ZN2at4_ops15to_dtype_layout4callERKNS_6TensorEN3c108optionalINS5_10ScalarTypeEEENS6_INS5_6LayoutEEENS6_INS5_6DeviceEEENS6_IbEEbbNS6_INS5_12MemoryFormatEEE
ok it seems this is sensitive to python version. You have to do 3.9. Then the link by rylan works.
Code:
conda create -n myenv python=3.9 -y
conda activate myenv
# needed in this order for vllm to work
pip install vllm
# optional
pip install -e ~/your_library/setup.py
ref: https://docs.vllm.ai/en/latest/getting_started/installation.html
@youkaichao what version of pytorch is supported then?
is it pytorch 2.1.2
?
If you build vllm from source, it supports (requires) pytorch 2.2 now.
If you build vllm from source, it supports (requires) pytorch 2.2 now.
I'm using pip though.
pip install vllm
directly, still uses pytorch 2.1 . stay tuned when we make the next release.
pip install vllm
directly, still uses pytorch 2.1 . stay tuned when we make the next release.
awesome! any estimate? Thank you!
@youkaichao since the versions of python I am using are fragile due to the current vllm (or one I use) only works with pytorch 2.1, I was wondering, what hugging face and accelerate version do we need without breaking vllm?
Need to debug this but I think this should work:
# for pytorch see doc string at the top of file
install_requires=[
'dill',
'networkx>=2.5',
'scipy',
'scikit-learn',
'lark-parser',
'tensorboard',
'pandas',
'progressbar2',
'requests',
'aiohttp',
'numpy',
'plotly',
'wandb',
'matplotlib',
# 'statsmodels'
# 'statsmodels==0.12.2'
# 'statsmodels==0.13.5'
# - later check why we are not installing it...
# 'seaborn'
# 'nltk'
'twine',
'torch==2.1.2', # 2.2 not supported due to vllm see: https://github.com/vllm-project/vllm/issues/2747
# 'torchvision',
# 'torchaudio',
# 'fairseq',
# 'trl',
'transformers==4.39.2', # my gold-ai-olympiad project uses 4.39.2
'accelerate==0.29.2',
# 'peft',
'datasets==2.18.0', # 2.18.0
'bitsandbytes== 0.43.0',
# 'einops',
'vllm==0.4.0.post1', # my gold-ai-olympiad project uses 0.4.0.post1 ref: https://github.com/vllm-project/vllm/issues/2747
]
)
and fyi:
(snap_cluster_setup) brando9@skampere1~/lean4ai $ pip list
Package Version
------------------------- ------------
absl-py 2.1.0
accelerate 0.29.2
aiohttp 3.9.3
aiosignal 1.3.1
annotated-types 0.6.0
anyio 4.3.0
appdirs 1.4.4
async-timeout 4.0.3
attrs 23.2.0
backports.tarfile 1.0.0
bitsandbytes 0.43.0
certifi 2024.2.2
cffi 1.16.0
charset-normalizer 3.3.2
click 8.1.7
cloudpickle 3.0.0
cmake 3.29.0.1
contourpy 1.2.1
cryptography 42.0.5
cycler 0.12.1
datasets 2.18.0
dill 0.3.8
diskcache 5.6.3
docker-pycreds 0.4.0
docutils 0.20.1
exceptiongroup 1.2.0
fastapi 0.110.1
filelock 3.13.3
fonttools 4.51.0
frozenlist 1.4.1
fsspec 2024.2.0
gitdb 4.0.11
GitPython 3.1.43
grpcio 1.62.1
h11 0.14.0
httptools 0.6.1
huggingface-hub 0.22.2
idna 3.6
importlib_metadata 7.1.0
importlib_resources 6.4.0
interegular 0.3.3
jaraco.classes 3.4.0
jaraco.context 5.3.0
jaraco.functools 4.0.0
jeepney 0.8.0
Jinja2 3.1.3
joblib 1.4.0
jsonschema 4.21.1
jsonschema-specifications 2023.12.1
keyring 25.1.0
kiwisolver 1.4.5
lark 1.1.9
lark-parser 0.12.0
llvmlite 0.42.0
Markdown 3.6
markdown-it-py 3.0.0
MarkupSafe 2.1.5
matplotlib 3.8.4
mdurl 0.1.2
more-itertools 10.2.0
mpmath 1.3.0
msgpack 1.0.8
multidict 6.0.5
multiprocess 0.70.16
nest-asyncio 1.6.0
networkx 3.2.1
nh3 0.2.17
ninja 1.11.1.1
numba 0.59.1
numpy 1.26.4
nvidia-cublas-cu12 12.1.3.1
nvidia-cuda-cupti-cu12 12.1.105
nvidia-cuda-nvrtc-cu12 12.1.105
nvidia-cuda-runtime-cu12 12.1.105
nvidia-cudnn-cu12 8.9.2.26
nvidia-cufft-cu12 11.0.2.54
nvidia-curand-cu12 10.3.2.106
nvidia-cusolver-cu12 11.4.5.107
nvidia-cusparse-cu12 12.1.0.106
nvidia-nccl-cu12 2.18.1
nvidia-nvjitlink-cu12 12.4.127
nvidia-nvtx-cu12 12.1.105
outlines 0.0.34
packaging 24.0
pandas 2.2.1
pillow 10.3.0
pip 24.0
pkginfo 1.10.0
plotly 5.20.0
progressbar2 4.4.2
prometheus_client 0.20.0
protobuf 4.25.3
psutil 5.9.8
py-cpuinfo 9.0.0
pyarrow 15.0.2
pyarrow-hotfix 0.6
pycparser 2.22
pydantic 2.6.4
pydantic_core 2.16.3
Pygments 2.17.2
pynvml 11.5.0
pyparsing 3.1.2
python-dateutil 2.9.0.post0
python-dotenv 1.0.1
python-utils 3.8.2
pytz 2024.1
PyYAML 6.0.1
ray 2.10.0
readme_renderer 43.0
referencing 0.34.0
regex 2023.12.25
requests 2.31.0
requests-toolbelt 1.0.0
rfc3986 2.0.0
rich 13.7.1
rpds-py 0.18.0
safetensors 0.4.2
scikit-learn 1.4.1.post1
scipy 1.13.0
SecretStorage 3.3.3
sentencepiece 0.2.0
sentry-sdk 1.44.1
setproctitle 1.3.3
setuptools 68.2.2
six 1.16.0
smmap 5.0.1
sniffio 1.3.1
starlette 0.37.2
sympy 1.12
tenacity 8.2.3
tensorboard 2.16.2
tensorboard-data-server 0.7.2
threadpoolctl 3.4.0
tiktoken 0.6.0
tokenizers 0.15.2
torch 2.1.2
tqdm 4.66.2
transformers 4.39.2
triton 2.1.0
twine 5.0.0
typing_extensions 4.11.0
tzdata 2024.1
urllib3 2.2.1
uvicorn 0.29.0
uvloop 0.19.0
vllm 0.4.0.post1
wandb 0.16.6
watchfiles 0.21.0
websockets 12.0
Werkzeug 3.0.2
wheel 0.41.2
xformers 0.0.23.post1
xxhash 3.4.1
yarl 1.9.4
zipp 3.18.1
For flash attention I have these comments
# # ampere
# 'dspy-ai',
# # 'torch==2.1.2+cu118', # 2.2 not supported due to vllm see: https://github.com/vllm-project/vllm/issues/2747
# # 'torch==2.1.2', # 2.2 not supported due to vllm see: https://github.com/vllm-project/vllm/issues/2747
# # 'torch==2.2.1', # 2.2 not supported due to vllm see: https://github.com/vllm-project/vllm/issues/2747
# 'torch==2.2.1', # 2.2 not supported due to vllm see: https://github.com/vllm-project/vllm/issues/2747
# # 'torchvision',
# # 'torchaudio',
# # 'trl',
# # 'transformers==4.39.2',
# 'transformers>=4.40',
# 'accelerate==0.29.2',
# # 'peft',
# # 'datasets==2.18.0',
# 'datasets==2.14.7',
# 'evaluate==0.4.1',
# 'bitsandbytes== 0.43.0',
# 'einops',
# 'flash-attn>=2.5.8',
# 'vllm==0.4.1', # my gold-ai-olympiad project uses 0.4.0.post1 ref: https://github.com/vllm-project/vllm/issues/2747
# # pip install -q -U google-generativeai
please install vllm in a fresh new environment, then you don't need to care about this manually.
I'm trying to run vllm and lm-eval-harness. I'm using vllm 0.2.5. After I'm done installing both, if I try importing vllm I get the following error:
File "/ramyapra/lm-evaluation-harness/lm_eval/models/__init__.py", line 7, in <module> from . import vllm_causallms File "/ramyapra/lm-evaluation-harness/lm_eval/models/vllm_causallms.py", line 16, in <module> from vllm import LLM, SamplingParams File "/ramyapra/vllm/vllm/__init__.py", line 3, in <module> from vllm.engine.arg_utils import AsyncEngineArgs, EngineArgs File "/ramyapra/vllm/vllm/engine/arg_utils.py", line 6, in <module> from vllm.config import (CacheConfig, ModelConfig, ParallelConfig, File "/ramyapra/vllm/vllm/config.py", line 9, in <module> from vllm.utils import get_cpu_memory, is_hip File "/ramyapra/vllm/vllm/utils.py", line 8, in <module> from vllm._C import cuda_utils ImportError: /ramyapra/vllm/vllm/_C.cpython-310-x86_64-linux-gnu.so: undefined symbol: _ZN2at4_ops19empty_memory_format4callEN3c108ArrayRefINS2_6SymIntEEESt8optionalINS2_10ScalarTypeEES6_INS2_6LayoutEES6_INS2_6DeviceEES6_IbES6_INS2_12MemoryFormatEE
I'm using the NGC docker container 23:10-py3.