OpenBMB / ToolBench

[ICLR'24 spotlight] An open platform for training, serving, and evaluating large language model for tool learning.
https://openbmb.github.io/ToolBench/
Apache License 2.0
4.6k stars 397 forks source link

CUDA Setup failed despite GPU being available. Please run the following command to get more information: #290

Open WuRui-Ella opened 3 days ago

WuRui-Ella commented 3 days ago

配置: 系统:CentOS Linux release 7.9.2009 CUDA版本: 12.4 python版本: 3.11 NVIDIA-SMI :550.54.14 我在使用hello_world推理时遇到了这个问题,已经提示重新编译了bitsandbytes,但还是出现了这个问题。

问题如下: CUDA SETUP: Something unexpected happened. Please compile from source: git clone git@github.com:TimDettmers/bitsandbytes.git cd bitsandbytes CUDA_VERSION=124 python setup.py install CUDA SETUP: Setup Failed! CUDA SETUP: Something unexpected happened. Please compile from source: git clone git@github.com:TimDettmers/bitsandbytes.git cd bitsandbytes CUDA_VERSION=124 python setup.py install ╭───────────────────── Traceback (most recent call last) ──────────────────────╮ │ /home/ToolBench-master/toolbench/inference/qa_pipeline.py:13 in │ │ │ │ 10 os.environ['CUDA_VERSION'] = "12.4" │ │ 11 os.environ['NVCC'] = '/usr/local/cuda-12.4/bin/nvcc' │ │ 12 │ │ ❱ 13 from toolbench.inference.Downstream_tasks.rapidapi import pipeline_runn │ │ 14 │ │ 15 │ │ 16 if name == "main": │ │ │ │ /home/ToolBench-master/toolbench/inference/Downstream_tasks/rapidapi.py:1 │ │ 0 in │ │ │ │ 7 from termcolor import colored │ │ 8 import random │ │ 9 from toolbench.inference.LLM.chatgpt_function_model import ChatGPTFunc │ │ ❱ 10 from toolbench.inference.LLM.davinci_model import Davinci │ │ 11 from toolbench.inference.LLM.tool_llama_lora_model import ToolLLaMALoR │ │ 12 from toolbench.inference.LLM.tool_llama_model import ToolLLaMA │ │ 13 from toolbench.inference.LLM.retriever import ToolRetriever │ │ │ │ /home/ToolBench-master/toolbench/inference/LLM/davinci_model.py:9 in │ │ │ │ │ │ 6 import random │ │ 7 import openai │ │ 8 from typing import Optional │ │ ❱ 9 from toolbench.model.model_adapter import get_conversation_template │ │ 10 from toolbench.inference.utils import SimpleChatIO, react_parser │ │ 11 from toolbench.inference.Prompts.ReAct_prompts import FORMAT_INSTRUCTI │ │ 12 │ │ │ │ /home/ToolBench-master/toolbench/model/init.py:1 in │ │ │ │ ❱ 1 from toolbench.model.model_adapter import ( │ │ 2 │ load_model, │ │ 3 │ get_conversation_template, │ │ 4 │ add_model_args, │ │ │ │ /home/ToolBench-master/toolbench/model/model_adapter.py:21 in │ │ │ │ 18 │ AutoTokenizer, │ │ 19 │ LlamaForCausalLM, │ │ 20 ) │ │ ❱ 21 from peft import PeftModel │ │ 22 │ │ 23 from toolbench.tool_conversation import Conversation, get_conv_templat │ │ 24 from toolbench.model.compression import load_compress_model │ │ │ │ /home/anaconda3/envs/toolllm/lib/python3.11/site-packages/peft/init.py:2 │ │ 2 in │ │ │ │ 19 │ │ 20 version = "0.3.0" │ │ 21 │ │ ❱ 22 from .mapping import MODEL_TYPE_TO_PEFT_MODEL_MAPPING, PEFT_TYPE_TO_CON │ │ 23 from .peft_model import ( │ │ 24 │ PeftModel, │ │ 25 │ PeftModelForCausalLM, │ │ │ │ /home/anaconda3/envs/toolllm/lib/python3.11/site-packages/peft/mapping.py:16 │ │ in │ │ │ │ 13 # See the License for the specific language governing permissions and │ │ 14 # limitations under the License. │ │ 15 │ │ ❱ 16 from .peft_model import ( │ │ 17 │ PeftModel, │ │ 18 │ PeftModelForCausalLM, │ │ 19 │ PeftModelForSeq2SeqLM, │ │ │ │ /home/anaconda3/envs/toolllm/lib/python3.11/site-packages/peft/peft_model.py │ │ :31 in │ │ │ │ 28 from transformers.modeling_outputs import SequenceClassifierOutput, T │ │ 29 from transformers.utils import PushToHubMixin │ │ 30 │ │ ❱ 31 from .tuners import ( │ │ 32 │ AdaLoraModel, │ │ 33 │ AdaptionPromptModel, │ │ 34 │ LoraModel, │ │ │ │ /home/anaconda3/envs/toolllm/lib/python3.11/site-packages/peft/tuners/init │ │ .py:21 in │ │ │ │ 18 # limitations under the License. │ │ 19 │ │ 20 from .adaption_prompt import AdaptionPromptConfig, AdaptionPromptModel │ │ ❱ 21 from .lora import LoraConfig, LoraModel │ │ 22 from .adalora import AdaLoraConfig, AdaLoraModel │ │ 23 from .p_tuning import PromptEncoder, PromptEncoderConfig, PromptEncoder │ │ 24 from .prefix_tuning import PrefixEncoder, PrefixTuningConfig │ │ │ │ /home/anaconda3/envs/toolllm/lib/python3.11/site-packages/peft/tuners/lora.p │ │ y:40 in │ │ │ │ 37 │ │ 38 │ │ 39 if is_bnb_available(): │ │ ❱ 40 │ import bitsandbytes as bnb │ │ 41 │ │ 42 │ │ 43 @dataclass │ │ │ │ /home/anaconda3/envs/toolllm/lib/python3.11/site-packages/bitsandbytes-0.38. │ │ 0.post2-py3.11.egg/bitsandbytes/init.py:7 in │ │ │ │ 4 # LICENSE file in the root directory of this source tree. │ │ 5 │ │ 6 from . import cuda_setup, utils │ │ ❱ 7 from .autograd._functions import ( │ │ 8 │ MatmulLtState, │ │ 9 │ bmm_cublas, │ │ 10 │ matmul, │ │ │ │ /home/anaconda3/envs/toolllm/lib/python3.11/site-packages/bitsandbytes-0.38. │ │ 0.post2-py3.11.egg/bitsandbytes/autograd/init.py:1 in │ │ │ │ ❱ 1 from ._functions import undo_layout, get_inverse_transform_indices │ │ 2 │ │ │ │ /home/anaconda3/envs/toolllm/lib/python3.11/site-packages/bitsandbytes-0.38. │ │ 0.post2-py3.11.egg/bitsandbytes/autograd/_functions.py:9 in │ │ │ │ 6 │ │ 7 import torch │ │ 8 │ │ ❱ 9 import bitsandbytes.functional as F │ │ 10 │ │ 11 │ │ 12 # math.prod not compatible with python < 3.8 │ │ │ │ /home/anaconda3/envs/toolllm/lib/python3.11/site-packages/bitsandbytes-0.38. │ │ 0.post2-py3.11.egg/bitsandbytes/functional.py:17 in │ │ │ │ 14 from typing import Tuple │ │ 15 from torch import Tensor │ │ 16 │ │ ❱ 17 from .cextension import COMPILED_WITH_CUDA, lib │ │ 18 │ │ 19 │ │ 20 # math.prod not compatible with python < 3.8 │ │ │ │ /home/anaconda3/envs/toolllm/lib/python3.11/site-packages/bitsandbytes-0.38. │ │ 0.post2-py3.11.egg/bitsandbytes/cextension.py:20 in │ │ │ │ 17 │ if lib is None and torch.cuda.is_available(): │ │ 18 │ │ CUDASetup.get_instance().generate_instructions() │ │ 19 │ │ CUDASetup.get_instance().print_log_stack() │ │ ❱ 20 │ │ raise RuntimeError(''' │ │ 21 │ │ CUDA Setup failed despite GPU being available. Please run the f │ │ 22 │ │ │ │ 23 │ │ python -m bitsandbytes │ ╰──────────────────────────────────────────────────────────────────────────────╯ RuntimeError: CUDA Setup failed despite GPU being available. Please run the following command to get more information:

    python -m bitsandbytes

    Inspect the output of the command and see if you can locate CUDA

libraries. You might need to add them to your LD_LIBRARY_PATH. If you suspect a bug, please take the information from python -m bitsandbytes and open an issue at: https://github.com/TimDettmers/bitsandbytes/issues ./scripts/inference_toolllama_helloworld_demo.sh:行3: --tool_root_dir: 未找到命令 ./scripts/inference_toolllama_helloworld_demo.sh:行4: --backbone_model: 未找到命令 ./scripts/inference_toolllama_helloworld_demo.sh:行5: --model_path: 未找到命令 ./scripts/inference_toolllama_helloworld_demo.sh:行6: --max_observation_length: 未找到命令 ./scripts/inference_toolllama_helloworld_demo.sh:行7: --observ_compress_method: 未找到命令 ./scripts/inference_toolllama_helloworld_demo.sh:行8: --method: 未找到命令 ./scripts/inference_toolllama_helloworld_demo.sh:行9: --input_query_file: 未找到命令 ./scripts/inference_toolllama_helloworld_demo.sh:行10: --output_answer_file: 未找到命令 ./scripts/inference_toolllama_helloworld_demo.sh:行11: --api_customization: 未找到命令

终端输入nvcc -V 微信图片_20240704153837 能看出来cuda是正确安装的,

找不到解决办法了55555555,特来问问大家,有知道怎么解决的嘛