hiyouga / LLaMA-Factory

Efficiently Fine-Tune 100+ LLMs in WebUI (ACL 2024)
https://arxiv.org/abs/2403.13372
Apache License 2.0
30.82k stars 3.81k forks source link

启动训练例子时报错,No such file or directory: tokenizer_config.json #4825

Closed rogerslh closed 2 months ago

rogerslh commented 2 months ago

Reminder

System Info

下载最新的版本 commit 99ab7a8c1c966232faa11b6a42b9740d9a20ace3 (grafted, HEAD -> main, origin/main, origin/HEAD) Author: hiyouga 467089858@qq.com Date: Mon Jul 15 01:16:26 2024 +0800

Reproduction

(venv) D:\LLaMA-Factory>llamafactory-cli train examples/train_lora/llama3_lora_sft.yaml [2024-07-15 11:44:52,402] [INFO] [real_accelerator.py:191:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-07-15 11:44:52,583] torch.distributed.elastic.multiprocessing.redirects: [WARNING] NOTE: Redirects are currently not supported in Windows or MacOs. 07/15/2024 11:44:56 - INFO - llamafactory.hparams.parser - Process rank: 0, device: cuda:0, n_gpu: 1, distributed training: False, compute dtype: torch.bfloat16 tokenizer_config.json: 100%|██████████████████████████████████████████████████████| 51.0k/51.0k [00:00<00:00, 19.4MB/s] Traceback (most recent call last): File "", line 198, in _run_module_as_main File "", line 88, in _run_code File "D:\unsloth_new\venv\Scripts\llamafactory-cli.exe__main__.py", line 7, in File "D:\LLaMA-Factory\src\llamafactory\cli.py", line 111, in main run_exp() File "D:\LLaMA-Factory\src\llamafactory\train\tuner.py", line 50, in run_exp run_sft(model_args, data_args, training_args, finetuning_args, generating_args, callbacks) File "D:\LLaMA-Factory\src\llamafactory\train\sft\workflow.py", line 44, in run_sft tokenizer_module = load_tokenizer(model_args) ^^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\LLaMA-Factory\src\llamafactory\model\loader.py", line 69, in load_tokenizer tokenizer = AutoTokenizer.from_pretrained( ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\unsloth_new\venv\Lib\site-packages\transformers\models\auto\tokenization_auto.py", line 826, in from_pretrained tokenizer_config = get_tokenizer_config(pretrained_model_name_or_path, **kwargs) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\unsloth_new\venv\Lib\site-packages\transformers\models\auto\tokenization_auto.py", line 679, in get_tokenizer_config with open(resolved_config_file, encoding="utf-8") as reader: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ FileNotFoundError: [Errno 2] No such file or directory: 'C:\Users\.cache\huggingface\hub\models--meta-llama--Meta-Llama-3-8B-Instruct\snapshots\e1945c40cd546c78e41f1151f4db032b271faeaa\tokenizer_config.json' Exception ignored in atexit callback: <function matmul_ext_update_autotune_table at 0x00000176FDB404A0> Traceback (most recent call last): File "D:\unsloth_new\venv\Lib\site-packages\deepspeed\ops\transformer\inference\triton\matmul_ext.py", line 444, in matmul_ext_update_autotune_table fp16_matmul._update_autotune_table() File "D:\unsloth_new\venv\Lib\site-packages\deepspeed\ops\transformer\inference\triton\matmul_ext.py", line 421, in _update_autotune_table TritonMatmul._update_autotune_table(class.name + "_2d_kernel", class._2d_kernel) File "D:\unsloth_new\venv\Lib\site-packages\deepspeed\ops\transformer\inference\triton\matmul_ext.py", line 150, in _update_autotune_table cache_manager.put(autotune_table) File "D:\unsloth_new\venv\Lib\site-packages\deepspeed\ops\transformer\inference\triton\matmul_ext.py", line 69, in put os.rename(self.file_path + ".tmp", self.file_path) FileExistsError: [WinError 183] 当文件已存在时,无法创建该文件。: 'C:\Users\.triton\autotune\Fp16Matmul_2d_kernel.pickle.tmp' -> 'C:\Users\.triton\autotune\Fp16Matmul_2d_kernel.pickle'

Expected behavior

No response

Others

No response

rogerslh commented 2 months ago

能告诉下出现这个问题的原因嘛