Open JV-X opened 2 months ago
Got some error. Hope authors to advice how to solve this problem:
2024-09-27 13:20:07 | INFO | model_worker | args: Namespace(host='0.0.0.0', port=40000, worker_address='http://localhost:40000', controller_address='http://localhost:10000', model_path='Llama-3.1-8B-Omni', model_base=None, model_name='Llama-3.1-8B-Omni', device='cuda', limit_model_concurrency=5, stream_interval=1, no_register=False, load_8bit=False, load_4bit=False, use_flash_attn=False, input_type='mel', mel_size=128, s2s=True, is_lora=False)
2024-09-27 13:20:09 | ERROR | stderr | Traceback (most recent call last):
2024-09-27 13:20:09 | ERROR | stderr | File "/home/user/anaconda3/envs/llama-omni/lib/python3.10/runpy.py", line 196, in _run_module_as_main
2024-09-27 13:20:09 | ERROR | stderr | return _run_code(code, main_globals, None,
2024-09-27 13:20:09 | ERROR | stderr | File "/home/user/anaconda3/envs/llama-omni/lib/python3.10/runpy.py", line 86, in _run_code
2024-09-27 13:20:09 | ERROR | stderr | exec(code, run_globals)
2024-09-27 13:20:09 | ERROR | stderr | File "/data1/git_proj/LLaMA-Omni/omni_speech/serve/model_worker.py", line 277, in
Got some error. Hope authors to advice how to solve this problem:
2024-09-27 13:20:07 | INFO | model_worker | args: Namespace(host='0.0.0.0', port=40000, worker_address='http://localhost:40000', controller_address='http://localhost:10000', model_path='Llama-3.1-8B-Omni', model_base=None, model_name='Llama-3.1-8B-Omni', device='cuda', limit_model_concurrency=5, stream_interval=1, no_register=False, load_8bit=False, load_4bit=False, use_flash_attn=False, input_type='mel', mel_size=128, s2s=True, is_lora=False) 2024-09-27 13:20:09 | ERROR | stderr | Traceback (most recent call last): 2024-09-27 13:20:09 | ERROR | stderr | File "/home/user/anaconda3/envs/llama-omni/lib/python3.10/runpy.py", line 196, in _run_module_as_main 2024-09-27 13:20:09 | ERROR | stderr | return _run_code(code, main_globals, None, 2024-09-27 13:20:09 | ERROR | stderr | File "/home/user/anaconda3/envs/llama-omni/lib/python3.10/runpy.py", line 86, in _run_code 2024-09-27 13:20:09 | ERROR | stderr | exec(code, run_globals) 2024-09-27 13:20:09 | ERROR | stderr | File "/data1/git_proj/LLaMA-Omni/omni_speech/serve/model_worker.py", line 277, in 2024-09-27 13:20:09 | ERROR | stderr | worker = ModelWorker(args.controller_address, 2024-09-27 13:20:09 | ERROR | stderr | File "/data1/git_proj/LLaMA-Omni/omni_speech/serve/model_worker.py", line 83, in init 2024-09-27 13:20:09 | ERROR | stderr | self.tokenizer, self.model, self.context_len = load_pretrained_model( 2024-09-27 13:20:09 | ERROR | stderr | File "/data1/git_proj/LLaMA-Omni/omni_speech/model/builder.py", line 79, in load_pretrained_model 2024-09-27 13:20:09 | ERROR | stderr | model = model_cls.from_pretrained( 2024-09-27 13:20:09 | ERROR | stderr | File "/home/user/anaconda3/envs/llama-omni/lib/python3.10/site-packages/transformers/modeling_utils.py", line 3798, in from_pretrained 2024-09-27 13:20:09 | ERROR | stderr | model = cls(config, *model_args, model_kwargs) 2024-09-27 13:20:09 | ERROR | stderr | File "/data1/git_proj/LLaMA-Omni/omni_speech/model/language_model/omni_speech2s_llama.py", line 25, in init 2024-09-27 13:20:09 | ERROR | stderr | super().init(config) 2024-09-27 13:20:09 | ERROR | stderr | File "/data1/git_proj/LLaMA-Omni/omni_speech/model/language_model/omni_speech_llama.py", line 46, in init 2024-09-27 13:20:09 | ERROR | stderr | self.model = OmniSpeechLlamaModel(config) 2024-09-27 13:20:09 | ERROR | stderr | File "/data1/git_proj/LLaMA-Omni/omni_speech/model/language_model/omni_speech_llama.py", line 38, in init 2024-09-27 13:20:09 | ERROR | stderr | super(OmniSpeechLlamaModel, self).init(config) 2024-09-27 13:20:09 | ERROR | stderr | File "/data1/git_proj/LLaMA-Omni/omni_speech/model/omni_speech_arch.py", line 32, in init 2024-09-27 13:20:09 | ERROR | stderr | self.speech_encoder = build_speech_encoder(config) 2024-09-27 13:20:09 | ERROR | stderr | File "/data1/git_proj/LLaMA-Omni/omni_speech/model/speech_encoder/builder.py", line 7, in build_speech_encoder 2024-09-27 13:20:09 | ERROR | stderr | return WhisperWrappedEncoder.load(config) 2024-09-27 13:20:09 | ERROR | stderr | File "/data1/git_proj/LLaMA-Omni/omni_speech/model/speech_encoder/speech_encoder.py", line 28, in load 2024-09-27 13:20:09 | ERROR | stderr | encoder = whisper.load_model(name=model_config.speech_encoder, device=device).encoder.to(dtype) 2024-09-27 13:20:09 | ERROR | stderr | File "/home/user/anaconda3/envs/llama-omni/lib/python3.10/site-packages/whisper/init.py", line 150, in load_model 2024-09-27 13:20:09 | ERROR | stderr | model = Whisper(dims) 2024-09-27 13:20:09 | ERROR | stderr | File "/home/user/anaconda3/envs/llama-omni/lib/python3.10/site-packages/whisper/model.py", line 225, in init 2024-09-27 13:20:09 | ERROR | stderr | self.encoder = AudioEncoder( 2024-09-27 13:20:09 | ERROR | stderr | File "/home/user/anaconda3/envs/llama-omni/lib/python3.10/site-packages/whisper/model.py", line 150, in init* 2024-09-27 13:20:09 | ERROR | stderr | self.register_buffer("positional_embedding", sinusoids(n_ctx, n_state)) 2024-09-27 13:20:09 | ERROR | stderr | File "/home/user/anaconda3/envs/llama-omni/lib/python3.10/site-packages/whisper/model.py", line 57, in sinusoids 2024-09-27 13:20:09 | ERROR | stderr | inv_timescales = torch.exp(-log_timescale_increment torch.arange(channels // 2)) 2024-09-27 13:20:09 | ERROR | stderr | RuntimeError: "exp_vml_cpu" not implemented for 'Half' (llama-omni) user@ HPC:/data1/git_proj/LLaMA-Omni$
Same here. I followed the set up in the readme and bash omni_speech/infer/run.sh omni_speech/infer/examples
but encounter this error
Traceback (most recent call last): File "/home/rczheng/LLaMA-Omni/omni_speech/infer/infer.py", line 181, in <module> eval_model(args) File "/home/rczheng/LLaMA-Omni/omni_speech/infer/infer.py", line 94, in eval_model tokenizer, model, context_len = load_pretrained_model(model_path, args.model_base, is_lora=args.is_lora, s2s=args.s2s) File "/home/rczheng/LLaMA-Omni/omni_speech/model/builder.py", line 79, in load_pretrained_model model = model_cls.from_pretrained( File "/home/rczheng/anaconda3/envs/llama-omni/lib/python3.10/site-packages/transformers/modeling_utils.py", line 3798, in from_pretrained model = cls(config, *model_args, **model_kwargs) File "/home/rczheng/LLaMA-Omni/omni_speech/model/language_model/omni_speech2s_llama.py", line 25, in __init__ super().__init__(config) File "/home/rczheng/LLaMA-Omni/omni_speech/model/language_model/omni_speech_llama.py", line 46, in __init__ self.model = OmniSpeechLlamaModel(config) File "/home/rczheng/LLaMA-Omni/omni_speech/model/language_model/omni_speech_llama.py", line 38, in __init__ super(OmniSpeechLlamaModel, self).__init__(config) File "/home/rczheng/LLaMA-Omni/omni_speech/model/omni_speech_arch.py", line 32, in __init__ self.speech_encoder = build_speech_encoder(config) File "/home/rczheng/LLaMA-Omni/omni_speech/model/speech_encoder/builder.py", line 7, in build_speech_encoder return WhisperWrappedEncoder.load(config) File "/home/rczheng/LLaMA-Omni/omni_speech/model/speech_encoder/speech_encoder.py", line 26, in load encoder = whisper.load_model(name=model_config.speech_encoder, device='cpu').encoder File "/home/rczheng/anaconda3/envs/llama-omni/lib/python3.10/site-packages/whisper/__init__.py", line 154, in load_model model = Whisper(dims) File "/home/rczheng/anaconda3/envs/llama-omni/lib/python3.10/site-packages/whisper/model.py", line 256, in __init__ self.encoder = AudioEncoder( File "/home/rczheng/anaconda3/envs/llama-omni/lib/python3.10/site-packages/whisper/model.py", line 181, in __init__ self.register_buffer("positional_embedding", sinusoids(n_ctx, n_state)) File "/home/rczheng/anaconda3/envs/llama-omni/lib/python3.10/site-packages/whisper/model.py", line 66, in sinusoids inv_timescales = torch.exp(-log_timescale_increment * torch.arange(channels // 2)) RuntimeError: "exp_vml_cpu" not implemented for 'Half'
I am following the steps in the readme.md document to install the environment on my Windows computer. When I execute
, the program automatically exits. What could be the reason? Could you help me take a look?
here is my error log