Open uni-manjunath-ke opened 3 months ago
But, not sure, what has to be given for ckpt_path, currently I do not have model.pt? Where do I get this? Is it some open-source model available in hugging face etc.? Please let me know. Currently, it is failing with this error. Thanks
You can find the projector(model.pt) ckpt in the README page.
Thanks @ddlBoJack .. Will try that
Thanks @ddlBoJack .. This is solved. But, I am facing another issue. I am running this in a 4 GPU machine. I got cuda OOM, when using only single GPU (through CUDA_VISIBLE_DEVICES="0" ). After this, I tried using all 4 GPU's by setting CUDA_VISIBLE_DEVICES="0,1,2,3", but I still get same cuda OOM.
I parallelly monitored nvidia-smi output, and found that the script is using only single GPU, in spite of setting CUDA_VISIBLE_DEVICES="0,1,2,3". Am, I missing anything here. Please suggest. Thanks
Stack trace: [2024-06-06 06:57:42][slam_model_asr.py][INFO] - loading other parts from: /mnt/efs/manju/if/repos/prompt/slam/output/vicuna-7b-v1.5-librispeech-linear-steplrwarmupkeep1e-4-wavlm-large-20240426/asr_epoch_1_step_1000/model.pt [2024-06-06 06:57:42][slam_llm.utils.train_utils][INFO] - --> Model asr [2024-06-06 06:57:42][slam_llm.utils.train_utils][INFO] - --> asr has 18.880512 Million params
Error executing job with overrides: ['++model_config.llm_name=vicuna-7b-v1.5', '++model_config.llm_path=lmsys/vicuna-7b-v1.5', '++model_config.llm_dim=4096', '++model_config.encoder_name=wavlm', '++model_config.normalize=true', '++dataset_config.normalize=true', '++model_config.encoder_projector_ds_rate=5', '++model_config.encoder_path=/mnt/efs/manju/if/repos/prompt/slam/models/WavLM-Large.pt', '++model_config.encoder_dim=1024', '++model_config.encoder_projector=linear', '++dataset_config.dataset=speech_dataset', '++dataset_config.val_data_path=/mnt/efs/manju/if/repos/prompt/slam/data/librispeech_slam_test-clean_bidisha.jsonl', '++dataset_config.input_type=raw', '++dataset_config.inference_mode=true', '++train_config.model_name=asr', '++train_config.freeze_encoder=true', '++train_config.freeze_llm=true', '++train_config.batching_strategy=custom', '++train_config.num_epochs=1', '++train_config.val_batch_size=1', '++train_config.num_workers_dataloader=2', '++train_config.output_dir=/mnt/efs/manju/if/repos/prompt/slam/output/vicuna-7b-v1.5-librispeech-linear-steplrwarmupkeep1e-4-wavlm-large-20240426', '++decode_log=/mnt/efs/manju/if/repos/prompt/slam/output/vicuna-7b-v1.5-librispeech-linear-steplrwarmupkeep1e-4-wavlm-large-20240426/asr_epoch_1_step_1000/decode_librispeech_test_clean_beam4', '++ckpt_path=/mnt/efs/manju/if/repos/prompt/slam/output/vicuna-7b-v1.5-librispeech-linear-steplrwarmupkeep1e-4-wavlm-large-20240426/asr_epoch_1_step_1000/model.pt']
Traceback (most recent call last):
File "/mnt/efs/manju/if/repos/prompt/slam/SLAM-LLM/examples/asr_librispeech/inference_asr_batch.py", line 53, in
Does this support multi-GPU? If not how can I resolve this issue. Thanks @ddlBoJack @LauraGPT @chenxie95
Currently we use a single GPU for decoding. We have the plan to support Multi-GPU decoding and the script is on the way.
Ok Thanks.. Does this mean, I should increase the GPU memory and run there? Thanks
What GPU do you use? You can set batch size to 1, and use half-precision to run the inference, if the GPU memory is limited.
We use nvidia a10g GPU with 24 GB GPU memory.. Please guide me on how to set these..
Currently, for batch size is set as " ++train_config.val_batch_size=1 " Thanks
Hi @ddlBoJack I tried several combinations by changing below parameters, but nothing works, & everything gives cuda OOM error. What is the minimum GPU needed to run these experiments. ++train_config.use_fp16=true \ ++train_config.use_peft=true \ ++train_config.one_gpu=true \ ++train_config.batch_size_training=1 \
And, Pls suggest, if there is any better way to make this work on GPU with 24 gb memory. Thanks
Hi @ddlBoJack @byrTony-Frankzyq Could you please suggest on this? Any recommendation on preferred GPU to be used ? Thanks
The minimum GPU we have used is A40 with 48GB memory.
Would be great if this is documented as either general requirement, or for each model combinations on minimum requirement (or experiments conducted on)
will be a great easy start! 🙂
Ok thank you @ddlBoJack . When can we expect multi gpu scripts availability? Any tentative ETA? Because, we have difficulty in accessing suggested high end GPU's .. so, we would want to try out in multi GPU environment.. thanks
Any update on this? @ddlBoJack @byrTony-Frankzyq Thanks
Hi, we can hardly provide an ETA, since all our contributors are part-time on the project. However, we will try our best to fix the existing bugs and claimed features by users.
@uni-manjunath-ke I am using the same spec of gpu. Could you let me know how did you run ./decode_wavlm_large_linear_vicuna_7b.sh
successfully?
Hi, I downloaded relevant models (encoders, projectors, llm) from the SLAM_LLM repo. And, placed in local folders, and updated these paths in ./decode_wavlm_large_linear_vicuna_7b.sh,
I built the slam_llm docker as given in Readme file and ran this script inside docker. It worked. Hope this will be useful.
@uni-manjunath-ke thank you for the reply! How did you do differently when you got OOM? When you didn't use docker, did you get OOM? I encountered OOM issue even with 'train config.use fp16=true'.
I used 40 GB single GPU with docker. With this, I dint get OOM
I used 4 46 GB GPUs. I still get an OOM error all the time, how should I fix this?
System Info
I am trying to run this: bash decode_wavlm_large_linear_vicuna_7b.sh
But, not sure, what has to be given for ckpt_path, currently I do not have model.pt? Where do I get this? Is it some open-source model available in hugging face etc.? Please let me know. Currently, it is failing with this error. Thanks @byrTony-Frankzyq
FileNotFoundError: [Errno 2] No such file or directory: '/mnt/efs/manju/if/repos/prompt/slam/output/vicuna-7b-v1.5-librispeech-linear-steplrwarmupkeep1e-4-wavlm-large-20240426/asr_epoch_1_step_1000/model.pt' Thanks
Information
🐛 Describe the bug
bash decode_wavlm_large_linear_vicuna_7b.sh
Error logs
FileNotFoundError: [Errno 2] No such file or directory: '/mnt/efs/manju/if/repos/prompt/slam/output/vicuna-7b-v1.5-librispeech-linear-steplrwarmupkeep1e-4-wavlm-large-20240426/asr_epoch_1_step_1000/model.pt'
Expected behavior
It is expected to produce output