liyucheng09 / MetaphorFrame

FrameBERT: Conceptual Metaphor Detection with Frame Embedding Learning. Presented at EACL 2023.
23 stars 2 forks source link

IndexError: index out of range in self #5

Open saishivakalyan opened 10 months ago

saishivakalyan commented 10 months ago

got this after running the latest run.sh.

bash-4.4$ ./scripts/run.sh /scratch/schalla7/Project/project_env/lib/python3.9/site-packages/torch/cuda/init.py:740: UserWarning: CUDA initialization: The NVIDIA driver on your system is too old (found version 11050). Please update your GPU driver by downloading and installing a new version from the URL: http://www.nvidia.com/Download/index.aspx Alternatively, go to: https://pytorch.org to install a PyTorch version that has been compiled with your version of the CUDA driver. (Triggered internally at ../c10/cuda/CUDAFunctions.cpp:108.) return torch._C._cuda_getDeviceCount() if nvml_count < 0 else nvml_count ['--data_dir', 'data/VUA18', '--task_name', 'vua', '--model_type', 'FrameMelbert', '--class_weight', '3', '--bert_model', 'roberta-base', '--num_train_epoch', '3', '--train_batch_size', '32', '--learning_rate', '3e-5', '--warmup_epoch', '2'] main config saved in /scratch/schalla7/Project/MetaphorFrame-main/checkpoints/framebert/roberta-base/2_20231117-2148 device: cpu n_gpu: 0 Some weights of RobertaModel were not initialized from the model checkpoint at roberta-base and are newly initialized: ['roberta.pooler.dense.bias', 'roberta.pooler.dense.weight'] You should probably TRAIN this model on a down-stream task to be able to use it for predictions and inference. Some weights of FrameFinder were not initialized from the model checkpoint at /scratch/schalla7/Project/MetaphorFrame-main/frame_finder/checkpoints/sent_no_mask_ff/ and are newly initialized: ['sent_classifier.weight', 'sent_classifier.bias'] You should probably TRAIN this model on a down-stream task to be able to use it for predictions and inference. Map: 100%|█████████████████████████████████████████████████████████████████████████████████████████████████| 19391/19391 [00:07<00:00, 2617.35 examples/s] Map: 100%|███████████████████████████████████████████████████████████████████████████████████████████████████| 2272/2272 [00:00<00:00, 2348.81 examples/s] Map: 100%|███████████████████████████████████████████████████████████████████████████████████████████████████| 6714/6714 [00:02<00:00, 2662.33 examples/s] /scratch/schalla7/Project/project_env/lib/python3.9/site-packages/transformers/optimization.py:411: FutureWarning: This implementation of AdamW is deprecated and will be removed in a future version. Use the PyTorch implementation torch.optim.AdamW instead, or set no_deprecation_warning=True to disable this warning warnings.warn( Running training Batch size = 32 Num steps = 9561 Epoch: 0%| You're using a RobertaTokenizerFast tokenizer. Please note that with a fast tokenizer, using the __call__ method is faster than using a method to encode the text followed by a call to the pad method to get a padded encoding. /scratch/schalla7/Project/project_env/lib/python3.9/site-packages/transformers/tokenization_utils_base.py:2624: UserWarning: max_length is ignored when padding=True and there is no truncation strategy. To pad to max length, use padding='max_length'. warnings.warn( Iteration: 0it [00:12, ?it/s] Epoch: 0%| | 0/3 [00:12<?, ?it/s] Traceback (most recent call last): File "/scratch/schalla7/Project/MetaphorFrame-main/main.py", line 602, in main() File "/scratch/schalla7/Project/MetaphorFrame-main/main.py", line 129, in main model, best_result = run_train( File "/scratch/schalla7/Project/MetaphorFrame-main/main.py", line 324, in run_train logits, frame_loss = model( File "/scratch/schalla7/Project/project_env/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1518, in _wrapped_call_impl return self._call_impl(*args, kwargs) File "/scratch/schalla7/Project/project_env/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1527, in _call_impl return forward_call(*args, *kwargs) File "/scratch/schalla7/Project/MetaphorFrame-main/modeling.py", line 649, in forward frame_task_output = self.frame_encoder(frame_input_ids, File "/scratch/schalla7/Project/project_env/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1518, in _wrapped_call_impl return self._call_impl(args, kwargs) File "/scratch/schalla7/Project/project_env/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1527, in _call_impl return forward_call(*args, kwargs) File "/scratch/schalla7/Project/MetaphorFrame-main/model.py", line 66, in forward outputs = self.roberta( File "/scratch/schalla7/Project/project_env/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1518, in _wrapped_call_impl return self._call_impl(*args, *kwargs) File "/scratch/schalla7/Project/project_env/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1527, in _call_impl return forward_call(args, kwargs) File "/scratch/schalla7/Project/project_env/lib/python3.9/site-packages/transformers/models/roberta/modeling_roberta.py", line 837, in forward embedding_output = self.embeddings( File "/scratch/schalla7/Project/project_env/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1518, in _wrapped_call_impl return self._call_impl(*args, kwargs) File "/scratch/schalla7/Project/project_env/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1527, in _call_impl return forward_call(*args, *kwargs) File "/scratch/schalla7/Project/project_env/lib/python3.9/site-packages/transformers/models/roberta/modeling_roberta.py", line 126, in forward token_type_embeddings = self.token_type_embeddings(token_type_ids) File "/scratch/schalla7/Project/project_env/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1518, in _wrapped_call_impl return self._call_impl(args, kwargs) File "/scratch/schalla7/Project/project_env/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1527, in _call_impl return forward_call(*args, **kwargs) File "/scratch/schalla7/Project/project_env/lib/python3.9/site-packages/torch/nn/modules/sparse.py", line 162, in forward return F.embedding( File "/scratch/schalla7/Project/project_env/lib/python3.9/site-packages/torch/nn/functional.py", line 2233, in embedding return torch.embedding(weight, input, padding_idx, scale_grad_by_freq, sparse) IndexError: index out of range in self

stops at line 324 In main.py.