Closed rohitpreddy07 closed 4 days ago
I believe your env does not correctly install the torch.
you can simply run "import torch" in your python env. then you can see the error.
Issue resolved it seems to be an issue with setting the environment variable.
However I am still unable to run the run_generation_gpu_woq_for_llama.py
script.
2024-07-05 17:35:38,878 - datasets - INFO - PyTorch version 2.1.0a0+git04048c2 available.
C:\Users\rohit\miniconda3\envs\llm\lib\site-packages\huggingface_hub\file_download.py:1132: FutureWarning: `resume_download` is deprecated and will be removed in version 1.0.0. Downloads always resume when possible. If you want to force a new download, use `force_download=True`.
warnings.warn(
config.json: 100%|██████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████| 911/911 [00:00<00:00, 181kB/s]
C:\Users\rohit\miniconda3\envs\llm\lib\site-packages\huggingface_hub\file_download.py:157: UserWarning: `huggingface_hub` cache-system uses symlinks by default to efficiently store duplicated files but your machine does not support them in C:\Users\rohit\.cache\huggingface\hub\models--Qwen--Qwen-7B-Chat. Caching files will still work but in a degraded version that might require more space on your disk. This warning can be disabled by setting the `HF_HUB_DISABLE_SYMLINKS_WARNING` environment variable. For more details, see https://huggingface.co/docs/huggingface_hub/how-to-cache#limitations.
To support symlinks on Windows, you either need to activate Developer Mode or to run Python as an administrator. In order to see activate developer mode, see this article: https://docs.microsoft.com/en-us/windows/apps/get-started/enable-your-device-for-development
warnings.warn(message)
Traceback (most recent call last):
File "C:\Users\rohit\intel-extension-for-pytorch\examples\gpu\inference\python\llm\run_generation_gpu_woq_for_llama.py", line 132, in <module>
config = AutoConfig.from_pretrained(
File "C:\Users\rohit\miniconda3\envs\llm\lib\site-packages\transformers\models\auto\configuration_auto.py", line 1051, in from_pretrained
trust_remote_code = resolve_trust_remote_code(
File "C:\Users\rohit\miniconda3\envs\llm\lib\site-packages\transformers\dynamic_module_utils.py", line 620, in resolve_trust_remote_code
raise ValueError(
ValueError: Loading Qwen/Qwen-7B-Chat requires you to execute the configuration file in that repo on your local machine. Make sure you have read the code there to avoid malicious use, then set the option `trust_remote_code=True` to remove this error.
Do you have an idea on which config file I have to execute to be able to run the script or am I misinterpreting the issue?
Issue resolved.
I am following the tutorial https://intel.github.io/intel-extension-for-pytorch/llm/llama3/xpu/ to run Llama 3 models locally, however I am getting the following error while setting up the environment and running the command
pip install -v .
.I'm not sure if its an error with pip as I ran
pip install --upgrade setuptools
, which was a popular solution upon researching this issue, with no avail. Please look into this issue.