bash-4.4$ ./scripts/run.sh
/scratch/schalla7/Project/project_env/lib/python3.9/site-packages/torch/cuda/init.py:740: UserWarning: CUDA initialization: The NVIDIA driver on your system is too old (found version 11050). Please update your GPU driver by downloading and installing a new version from the URL: http://www.nvidia.com/Download/index.aspx Alternatively, go to: https://pytorch.org to install a PyTorch version that has been compiled with your version of the CUDA driver. (Triggered internally at ../c10/cuda/CUDAFunctions.cpp:108.)
return torch._C._cuda_getDeviceCount() if nvml_count < 0 else nvml_count
['--data_dir', 'data/VUA18', '--task_name', 'vua', '--model_type', 'FrameMelbert', '--class_weight', '3', '--bert_model', 'roberta-base', '--num_train_epoch', '3', '--train_batch_size', '32', '--learning_rate', '3e-5', '--warmup_epoch', '2']
main config saved in /scratch/schalla7/Project/MetaphorFrame-main/checkpoints/framebert/roberta-base/2_20231117-2148
device: cpu n_gpu: 0
Some weights of RobertaModel were not initialized from the model checkpoint at roberta-base and are newly initialized: ['roberta.pooler.dense.bias', 'roberta.pooler.dense.weight']
You should probably TRAIN this model on a down-stream task to be able to use it for predictions and inference.
Some weights of FrameFinder were not initialized from the model checkpoint at /scratch/schalla7/Project/MetaphorFrame-main/frame_finder/checkpoints/sent_no_mask_ff/ and are newly initialized: ['sent_classifier.weight', 'sent_classifier.bias']
You should probably TRAIN this model on a down-stream task to be able to use it for predictions and inference.
Map: 100%|█████████████████████████████████████████████████████████████████████████████████████████████████| 19391/19391 [00:07<00:00, 2617.35 examples/s]
Map: 100%|███████████████████████████████████████████████████████████████████████████████████████████████████| 2272/2272 [00:00<00:00, 2348.81 examples/s]
Map: 100%|███████████████████████████████████████████████████████████████████████████████████████████████████| 6714/6714 [00:02<00:00, 2662.33 examples/s]
/scratch/schalla7/Project/project_env/lib/python3.9/site-packages/transformers/optimization.py:411: FutureWarning: This implementation of AdamW is deprecated and will be removed in a future version. Use the PyTorch implementation torch.optim.AdamW instead, or set no_deprecation_warning=True to disable this warning
warnings.warn(
Running training
Batch size = 32
Num steps = 9561
Epoch: 0%| You're using a RobertaTokenizerFast tokenizer. Please note that with a fast tokenizer, using the __call__ method is faster than using a method to encode the text followed by a call to the pad method to get a padded encoding.
/scratch/schalla7/Project/project_env/lib/python3.9/site-packages/transformers/tokenization_utils_base.py:2624: UserWarning: max_length is ignored when padding=True and there is no truncation strategy. To pad to max length, use padding='max_length'.
warnings.warn(
Iteration: 0it [00:12, ?it/s]
Epoch: 0%| | 0/3 [00:12<?, ?it/s]
Traceback (most recent call last):
File "/scratch/schalla7/Project/MetaphorFrame-main/main.py", line 602, in
main()
File "/scratch/schalla7/Project/MetaphorFrame-main/main.py", line 129, in main
model, best_result = run_train(
File "/scratch/schalla7/Project/MetaphorFrame-main/main.py", line 324, in run_train
logits, frame_loss = model(
File "/scratch/schalla7/Project/project_env/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1518, in _wrapped_call_impl
return self._call_impl(*args, kwargs)
File "/scratch/schalla7/Project/project_env/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1527, in _call_impl
return forward_call(*args, *kwargs)
File "/scratch/schalla7/Project/MetaphorFrame-main/modeling.py", line 649, in forward
frame_task_output = self.frame_encoder(frame_input_ids,
File "/scratch/schalla7/Project/project_env/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1518, in _wrapped_call_impl
return self._call_impl(args, kwargs)
File "/scratch/schalla7/Project/project_env/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1527, in _call_impl
return forward_call(*args, kwargs)
File "/scratch/schalla7/Project/MetaphorFrame-main/model.py", line 66, in forward
outputs = self.roberta(
File "/scratch/schalla7/Project/project_env/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1518, in _wrapped_call_impl
return self._call_impl(*args, *kwargs)
File "/scratch/schalla7/Project/project_env/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1527, in _call_impl
return forward_call(args, kwargs)
File "/scratch/schalla7/Project/project_env/lib/python3.9/site-packages/transformers/models/roberta/modeling_roberta.py", line 837, in forward
embedding_output = self.embeddings(
File "/scratch/schalla7/Project/project_env/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1518, in _wrapped_call_impl
return self._call_impl(*args, kwargs)
File "/scratch/schalla7/Project/project_env/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1527, in _call_impl
return forward_call(*args, *kwargs)
File "/scratch/schalla7/Project/project_env/lib/python3.9/site-packages/transformers/models/roberta/modeling_roberta.py", line 126, in forward
token_type_embeddings = self.token_type_embeddings(token_type_ids)
File "/scratch/schalla7/Project/project_env/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1518, in _wrapped_call_impl
return self._call_impl(args, kwargs)
File "/scratch/schalla7/Project/project_env/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1527, in _call_impl
return forward_call(*args, **kwargs)
File "/scratch/schalla7/Project/project_env/lib/python3.9/site-packages/torch/nn/modules/sparse.py", line 162, in forward
return F.embedding(
File "/scratch/schalla7/Project/project_env/lib/python3.9/site-packages/torch/nn/functional.py", line 2233, in embedding
return torch.embedding(weight, input, padding_idx, scale_grad_by_freq, sparse)
IndexError: index out of range in self
got this after running the latest run.sh.
bash-4.4$ ./scripts/run.sh /scratch/schalla7/Project/project_env/lib/python3.9/site-packages/torch/cuda/init.py:740: UserWarning: CUDA initialization: The NVIDIA driver on your system is too old (found version 11050). Please update your GPU driver by downloading and installing a new version from the URL: http://www.nvidia.com/Download/index.aspx Alternatively, go to: https://pytorch.org to install a PyTorch version that has been compiled with your version of the CUDA driver. (Triggered internally at ../c10/cuda/CUDAFunctions.cpp:108.) return torch._C._cuda_getDeviceCount() if nvml_count < 0 else nvml_count ['--data_dir', 'data/VUA18', '--task_name', 'vua', '--model_type', 'FrameMelbert', '--class_weight', '3', '--bert_model', 'roberta-base', '--num_train_epoch', '3', '--train_batch_size', '32', '--learning_rate', '3e-5', '--warmup_epoch', '2'] main config saved in /scratch/schalla7/Project/MetaphorFrame-main/checkpoints/framebert/roberta-base/2_20231117-2148 device: cpu n_gpu: 0 Some weights of RobertaModel were not initialized from the model checkpoint at roberta-base and are newly initialized: ['roberta.pooler.dense.bias', 'roberta.pooler.dense.weight'] You should probably TRAIN this model on a down-stream task to be able to use it for predictions and inference. Some weights of FrameFinder were not initialized from the model checkpoint at /scratch/schalla7/Project/MetaphorFrame-main/frame_finder/checkpoints/sent_no_mask_ff/ and are newly initialized: ['sent_classifier.weight', 'sent_classifier.bias'] You should probably TRAIN this model on a down-stream task to be able to use it for predictions and inference. Map: 100%|█████████████████████████████████████████████████████████████████████████████████████████████████| 19391/19391 [00:07<00:00, 2617.35 examples/s] Map: 100%|███████████████████████████████████████████████████████████████████████████████████████████████████| 2272/2272 [00:00<00:00, 2348.81 examples/s] Map: 100%|███████████████████████████████████████████████████████████████████████████████████████████████████| 6714/6714 [00:02<00:00, 2662.33 examples/s] /scratch/schalla7/Project/project_env/lib/python3.9/site-packages/transformers/optimization.py:411: FutureWarning: This implementation of AdamW is deprecated and will be removed in a future version. Use the PyTorch implementation torch.optim.AdamW instead, or set
main()
File "/scratch/schalla7/Project/MetaphorFrame-main/main.py", line 129, in main
model, best_result = run_train(
File "/scratch/schalla7/Project/MetaphorFrame-main/main.py", line 324, in run_train
logits, frame_loss = model(
File "/scratch/schalla7/Project/project_env/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1518, in _wrapped_call_impl
return self._call_impl(*args, kwargs)
File "/scratch/schalla7/Project/project_env/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1527, in _call_impl
return forward_call(*args, *kwargs)
File "/scratch/schalla7/Project/MetaphorFrame-main/modeling.py", line 649, in forward
frame_task_output = self.frame_encoder(frame_input_ids,
File "/scratch/schalla7/Project/project_env/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1518, in _wrapped_call_impl
return self._call_impl(args, kwargs)
File "/scratch/schalla7/Project/project_env/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1527, in _call_impl
return forward_call(*args, kwargs)
File "/scratch/schalla7/Project/MetaphorFrame-main/model.py", line 66, in forward
outputs = self.roberta(
File "/scratch/schalla7/Project/project_env/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1518, in _wrapped_call_impl
return self._call_impl(*args, *kwargs)
File "/scratch/schalla7/Project/project_env/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1527, in _call_impl
return forward_call(args, kwargs)
File "/scratch/schalla7/Project/project_env/lib/python3.9/site-packages/transformers/models/roberta/modeling_roberta.py", line 837, in forward
embedding_output = self.embeddings(
File "/scratch/schalla7/Project/project_env/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1518, in _wrapped_call_impl
return self._call_impl(*args, kwargs)
File "/scratch/schalla7/Project/project_env/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1527, in _call_impl
return forward_call(*args, *kwargs)
File "/scratch/schalla7/Project/project_env/lib/python3.9/site-packages/transformers/models/roberta/modeling_roberta.py", line 126, in forward
token_type_embeddings = self.token_type_embeddings(token_type_ids)
File "/scratch/schalla7/Project/project_env/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1518, in _wrapped_call_impl
return self._call_impl(args, kwargs)
File "/scratch/schalla7/Project/project_env/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1527, in _call_impl
return forward_call(*args, **kwargs)
File "/scratch/schalla7/Project/project_env/lib/python3.9/site-packages/torch/nn/modules/sparse.py", line 162, in forward
return F.embedding(
File "/scratch/schalla7/Project/project_env/lib/python3.9/site-packages/torch/nn/functional.py", line 2233, in embedding
return torch.embedding(weight, input, padding_idx, scale_grad_by_freq, sparse)
IndexError: index out of range in self
no_deprecation_warning=True
to disable this warning warnings.warn( Running training Batch size = 32 Num steps = 9561 Epoch: 0%| You're using a RobertaTokenizerFast tokenizer. Please note that with a fast tokenizer, using the__call__
method is faster than using a method to encode the text followed by a call to thepad
method to get a padded encoding. /scratch/schalla7/Project/project_env/lib/python3.9/site-packages/transformers/tokenization_utils_base.py:2624: UserWarning:max_length
is ignored whenpadding
=True
and there is no truncation strategy. To pad to max length, usepadding='max_length'
. warnings.warn( Iteration: 0it [00:12, ?it/s] Epoch: 0%| | 0/3 [00:12<?, ?it/s] Traceback (most recent call last): File "/scratch/schalla7/Project/MetaphorFrame-main/main.py", line 602, instops at line 324 In main.py.