(venv) D:\LLaMA-Factory>llamafactory-cli train examples/train_lora/llama3_lora_sft.yaml
[2024-07-15 11:44:52,402] [INFO] [real_accelerator.py:191:get_accelerator] Setting ds_accelerator to cuda (auto detect)
[2024-07-15 11:44:52,583] torch.distributed.elastic.multiprocessing.redirects: [WARNING] NOTE: Redirects are currently not supported in Windows or MacOs.
07/15/2024 11:44:56 - INFO - llamafactory.hparams.parser - Process rank: 0, device: cuda:0, n_gpu: 1, distributed training: False, compute dtype: torch.bfloat16
tokenizer_config.json: 100%|██████████████████████████████████████████████████████| 51.0k/51.0k [00:00<00:00, 19.4MB/s]
Traceback (most recent call last):
File "", line 198, in _run_module_as_main
File "", line 88, in _run_code
File "D:\unsloth_new\venv\Scripts\llamafactory-cli.exe__main__.py", line 7, in
File "D:\LLaMA-Factory\src\llamafactory\cli.py", line 111, in main
run_exp()
File "D:\LLaMA-Factory\src\llamafactory\train\tuner.py", line 50, in run_exp
run_sft(model_args, data_args, training_args, finetuning_args, generating_args, callbacks)
File "D:\LLaMA-Factory\src\llamafactory\train\sft\workflow.py", line 44, in run_sft
tokenizer_module = load_tokenizer(model_args)
^^^^^^^^^^^^^^^^^^^^^^^^^^
File "D:\LLaMA-Factory\src\llamafactory\model\loader.py", line 69, in load_tokenizer
tokenizer = AutoTokenizer.from_pretrained(
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "D:\unsloth_new\venv\Lib\site-packages\transformers\models\auto\tokenization_auto.py", line 826, in from_pretrained
tokenizer_config = get_tokenizer_config(pretrained_model_name_or_path, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "D:\unsloth_new\venv\Lib\site-packages\transformers\models\auto\tokenization_auto.py", line 679, in get_tokenizer_config
with open(resolved_config_file, encoding="utf-8") as reader:
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
FileNotFoundError: [Errno 2] No such file or directory: 'C:\Users\.cache\huggingface\hub\models--meta-llama--Meta-Llama-3-8B-Instruct\snapshots\e1945c40cd546c78e41f1151f4db032b271faeaa\tokenizer_config.json'
Exception ignored in atexit callback: <function matmul_ext_update_autotune_table at 0x00000176FDB404A0>
Traceback (most recent call last):
File "D:\unsloth_new\venv\Lib\site-packages\deepspeed\ops\transformer\inference\triton\matmul_ext.py", line 444, in matmul_ext_update_autotune_table
fp16_matmul._update_autotune_table()
File "D:\unsloth_new\venv\Lib\site-packages\deepspeed\ops\transformer\inference\triton\matmul_ext.py", line 421, in _update_autotune_table
TritonMatmul._update_autotune_table(class.name + "_2d_kernel", class._2d_kernel)
File "D:\unsloth_new\venv\Lib\site-packages\deepspeed\ops\transformer\inference\triton\matmul_ext.py", line 150, in _update_autotune_table
cache_manager.put(autotune_table)
File "D:\unsloth_new\venv\Lib\site-packages\deepspeed\ops\transformer\inference\triton\matmul_ext.py", line 69, in put
os.rename(self.file_path + ".tmp", self.file_path)
FileExistsError: [WinError 183] 当文件已存在时,无法创建该文件。: 'C:\Users\.triton\autotune\Fp16Matmul_2d_kernel.pickle.tmp' -> 'C:\Users\.triton\autotune\Fp16Matmul_2d_kernel.pickle'
Reminder
System Info
下载最新的版本 commit 99ab7a8c1c966232faa11b6a42b9740d9a20ace3 (grafted, HEAD -> main, origin/main, origin/HEAD) Author: hiyouga 467089858@qq.com Date: Mon Jul 15 01:16:26 2024 +0800
Reproduction
(venv) D:\LLaMA-Factory>llamafactory-cli train examples/train_lora/llama3_lora_sft.yaml [2024-07-15 11:44:52,402] [INFO] [real_accelerator.py:191:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-07-15 11:44:52,583] torch.distributed.elastic.multiprocessing.redirects: [WARNING] NOTE: Redirects are currently not supported in Windows or MacOs. 07/15/2024 11:44:56 - INFO - llamafactory.hparams.parser - Process rank: 0, device: cuda:0, n_gpu: 1, distributed training: False, compute dtype: torch.bfloat16 tokenizer_config.json: 100%|██████████████████████████████████████████████████████| 51.0k/51.0k [00:00<00:00, 19.4MB/s] Traceback (most recent call last): File "", line 198, in _run_module_as_main
File "", line 88, in _run_code
File "D:\unsloth_new\venv\Scripts\llamafactory-cli.exe__main__.py", line 7, in
File "D:\LLaMA-Factory\src\llamafactory\cli.py", line 111, in main
run_exp()
File "D:\LLaMA-Factory\src\llamafactory\train\tuner.py", line 50, in run_exp
run_sft(model_args, data_args, training_args, finetuning_args, generating_args, callbacks)
File "D:\LLaMA-Factory\src\llamafactory\train\sft\workflow.py", line 44, in run_sft
tokenizer_module = load_tokenizer(model_args)
^^^^^^^^^^^^^^^^^^^^^^^^^^
File "D:\LLaMA-Factory\src\llamafactory\model\loader.py", line 69, in load_tokenizer
tokenizer = AutoTokenizer.from_pretrained(
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "D:\unsloth_new\venv\Lib\site-packages\transformers\models\auto\tokenization_auto.py", line 826, in from_pretrained
tokenizer_config = get_tokenizer_config(pretrained_model_name_or_path, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "D:\unsloth_new\venv\Lib\site-packages\transformers\models\auto\tokenization_auto.py", line 679, in get_tokenizer_config
with open(resolved_config_file, encoding="utf-8") as reader:
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
FileNotFoundError: [Errno 2] No such file or directory: 'C:\Users\.cache\huggingface\hub\models--meta-llama--Meta-Llama-3-8B-Instruct\snapshots\e1945c40cd546c78e41f1151f4db032b271faeaa\tokenizer_config.json'
Exception ignored in atexit callback: <function matmul_ext_update_autotune_table at 0x00000176FDB404A0>
Traceback (most recent call last):
File "D:\unsloth_new\venv\Lib\site-packages\deepspeed\ops\transformer\inference\triton\matmul_ext.py", line 444, in matmul_ext_update_autotune_table
fp16_matmul._update_autotune_table()
File "D:\unsloth_new\venv\Lib\site-packages\deepspeed\ops\transformer\inference\triton\matmul_ext.py", line 421, in _update_autotune_table
TritonMatmul._update_autotune_table(class.name + "_2d_kernel", class._2d_kernel)
File "D:\unsloth_new\venv\Lib\site-packages\deepspeed\ops\transformer\inference\triton\matmul_ext.py", line 150, in _update_autotune_table
cache_manager.put(autotune_table)
File "D:\unsloth_new\venv\Lib\site-packages\deepspeed\ops\transformer\inference\triton\matmul_ext.py", line 69, in put
os.rename(self.file_path + ".tmp", self.file_path)
FileExistsError: [WinError 183] 当文件已存在时,无法创建该文件。: 'C:\Users\.triton\autotune\Fp16Matmul_2d_kernel.pickle.tmp' -> 'C:\Users\.triton\autotune\Fp16Matmul_2d_kernel.pickle'
Expected behavior
No response
Others
No response