Closed mikeleatila closed 1 year ago
@oroszgy @mfa @justindujardin @luckytoilet
@mikeleatila, were you able to solve this? I am getting this error now!
This issue is being closed due to lack of activity. If you think it still needs to be addressed, please comment on this thread 👇
Hello, I am seeing this error - could you please help/advise how to fix it? Many thanks in advance!
/home/mikeleatila/anaconda3/envs/domains/bin/python /home/mikeleatila/dont_stop_pretraining_master/scripts/train.py --config training_config/classifier.jsonnet --serialization_dir model_logs/citation-intent-dapt-dapt --hyperparameters ROBERTA_CLASSIFIER_SMALL --dataset citation_intent --model /home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688 --device 0 --perf +f1 --evaluate_on_test 2022-11-29 20:09:53,526 - INFO - transformers.file_utils - PyTorch version 1.13.0 available. 2022-11-29 20:09:54,138 - INFO - pytorch_pretrained_bert.modeling - Better speed can be achieved with apex installed from https://www.github.com/nvidia/apex . 2022-11-29 20:09:54,170 - INFO - allennlp.common.params - random_seed = 237386 2022-11-29 20:09:54,170 - INFO - allennlp.common.params - numpy_seed = 237386 2022-11-29 20:09:54,170 - INFO - allennlp.common.params - pytorch_seed = 237386 2022-11-29 20:09:54,171 - INFO - allennlp.common.checks - Pytorch version: 1.13.0 2022-11-29 20:09:54,171 - INFO - allennlp.common.from_params - instantiating class <class 'allennlp.commands.train.TrainModel'> from params {'trainer': {'cuda_device': 0, 'gradient_accumulation_batch_size': 16, 'num_epochs': 10, 'num_serialized_models_to_keep': 0, 'optimizer': {'b1': 0.9, 'b2': 0.98, 'e': 1e-06, 'lr': '2e-05', 'max_grad_norm': 1, 'parameter_groups': [[['bias', 'LayerNorm.bias', 'LayerNorm.weight', 'layer_norm.weight'], {'weight_decay': 0}, []]], 'schedule': 'warmup_linear', 't_total': -1, 'type': 'bert_adam', 'warmup': 0.06, 'weight_decay': 0.1}, 'patience': 3, 'validation_metric': '+f1'}, 'model': {'dropout': '0.1', 'feedforward_layer': {'activations': 'tanh', 'hidden_dims': 768, 'input_dim': 768, 'num_layers': 1}, 'seq2vec_encoder': {'embedding_dim': 768, 'type': 'cls_pooler_x'}, 'text_field_embedder': {'roberta': {'model_name': '/home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688', 'type': 'pretrained_transformer'}}, 'type': 'basic_classifier_with_f1'}, 'dataset_reader': {'lazy': False, 'max_sequence_length': 512, 'token_indexers': {'roberta': {'model_name': '/home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688', 'type': 'pretrained_transformer'}}, 'tokenizer': {'model_name': '/home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688', 'type': 'pretrained_transformer'}, 'type': 'text_classification_json_with_sampling'}, 'validation_data_path': 'https://s3-us-west-2.amazonaws.com/allennlp/dont_stop_pretraining/data/citation_intent/dev.jsonl', 'evaluate_on_test': True, 'iterator': {'batch_size': 16, 'sorting_keys': [['tokens', 'num_tokens']], 'type': 'bucket'}, 'test_data_path': 'https://s3-us-west-2.amazonaws.com/allennlp/dont_stop_pretraining/data/citation_intent/test.jsonl', 'validation_iterator': {'batch_size': 64, 'sorting_keys': [['tokens', 'num_tokens']], 'type': 'bucket'}, 'validation_dataset_reader': {'lazy': False, 'max_sequence_length': 512, 'token_indexers': {'roberta': {'model_name': '/home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688', 'type': 'pretrained_transformer'}}, 'tokenizer': {'model_name': '/home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688', 'type': 'pretrained_transformer'}, 'type': 'text_classification_json_with_sampling'}, 'train_data_path': 'https://s3-us-west-2.amazonaws.com/allennlp/dont_stop_pretraining/data/citation_intent/train.jsonl'} and extras {'local_rank', 'serialization_dir', 'batch_weight_key'} 2022-11-29 20:09:54,171 - INFO - allennlp.common.params - type = default 2022-11-29 20:09:54,171 - INFO - allennlp.common.from_params - instantiating class <class 'allennlp.commands.train.TrainModel'> from params {'trainer': {'cuda_device': 0, 'gradient_accumulation_batch_size': 16, 'num_epochs': 10, 'num_serialized_models_to_keep': 0, 'optimizer': {'b1': 0.9, 'b2': 0.98, 'e': 1e-06, 'lr': '2e-05', 'max_grad_norm': 1, 'parameter_groups': [[['bias', 'LayerNorm.bias', 'LayerNorm.weight', 'layer_norm.weight'], {'weight_decay': 0}, []]], 'schedule': 'warmup_linear', 't_total': -1, 'type': 'bert_adam', 'warmup': 0.06, 'weight_decay': 0.1}, 'patience': 3, 'validation_metric': '+f1'}, 'model': {'dropout': '0.1', 'feedforward_layer': {'activations': 'tanh', 'hidden_dims': 768, 'input_dim': 768, 'num_layers': 1}, 'seq2vec_encoder': {'embedding_dim': 768, 'type': 'cls_pooler_x'}, 'text_field_embedder': {'roberta': {'model_name': '/home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688', 'type': 'pretrained_transformer'}}, 'type': 'basic_classifier_with_f1'}, 'dataset_reader': {'lazy': False, 'max_sequence_length': 512, 'token_indexers': {'roberta': {'model_name': '/home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688', 'type': 'pretrained_transformer'}}, 'tokenizer': {'model_name': '/home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688', 'type': 'pretrained_transformer'}, 'type': 'text_classification_json_with_sampling'}, 'validation_data_path': 'https://s3-us-west-2.amazonaws.com/allennlp/dont_stop_pretraining/data/citation_intent/dev.jsonl', 'evaluate_on_test': True, 'iterator': {'batch_size': 16, 'sorting_keys': [['tokens', 'num_tokens']], 'type': 'bucket'}, 'test_data_path': 'https://s3-us-west-2.amazonaws.com/allennlp/dont_stop_pretraining/data/citation_intent/test.jsonl', 'validation_iterator': {'batch_size': 64, 'sorting_keys': [['tokens', 'num_tokens']], 'type': 'bucket'}, 'validation_dataset_reader': {'lazy': False, 'max_sequence_length': 512, 'token_indexers': {'roberta': {'model_name': '/home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688', 'type': 'pretrained_transformer'}}, 'tokenizer': {'model_name': '/home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688', 'type': 'pretrained_transformer'}, 'type': 'text_classification_json_with_sampling'}, 'train_data_path': 'https://s3-us-west-2.amazonaws.com/allennlp/dont_stop_pretraining/data/citation_intent/train.jsonl'} and extras {'local_rank', 'serialization_dir', 'batch_weight_key'} 2022-11-29 20:09:54,171 - INFO - allennlp.common.from_params - instantiating class <class 'allennlp.data.dataset_readers.dataset_reader.DatasetReader'> from params {'lazy': False, 'max_sequence_length': 512, 'token_indexers': {'roberta': {'model_name': '/home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688', 'type': 'pretrained_transformer'}}, 'tokenizer': {'model_name': '/home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688', 'type': 'pretrained_transformer'}, 'type': 'text_classification_json_with_sampling'} and extras {'local_rank', 'serialization_dir', 'batch_weight_key'} 2022-11-29 20:09:54,172 - INFO - allennlp.common.params - dataset_reader.type = text_classification_json_with_sampling 2022-11-29 20:09:54,172 - INFO - allennlp.common.from_params - instantiating class <class 'dont_stop_pretraining.data.dataset_readers.text_classification_json_reader_with_sampling.TextClassificationJsonReaderWithSampling'> from params {'lazy': False, 'max_sequence_length': 512, 'token_indexers': {'roberta': {'model_name': '/home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688', 'type': 'pretrained_transformer'}}, 'tokenizer': {'model_name': '/home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688', 'type': 'pretrained_transformer'}} and extras {'local_rank', 'serialization_dir', 'batch_weight_key'} 2022-11-29 20:09:54,172 - INFO - allennlp.common.from_params - instantiating class <class 'allennlp.data.token_indexers.token_indexer.TokenIndexer'> from params {'model_name': '/home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688', 'type': 'pretrained_transformer'} and extras {'local_rank', 'serialization_dir', 'batch_weight_key'} 2022-11-29 20:09:54,172 - INFO - allennlp.common.params - dataset_reader.token_indexers.roberta.type = pretrained_transformer 2022-11-29 20:09:54,172 - INFO - allennlp.common.from_params - instantiating class <class 'allennlp.data.token_indexers.pretrained_transformer_indexer.PretrainedTransformerIndexer'> from params {'model_name': '/home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688'} and extras {'local_rank', 'serialization_dir', 'batch_weight_key'} 2022-11-29 20:09:54,172 - INFO - allennlp.common.params - dataset_reader.token_indexers.roberta.token_min_padding_length = 0 2022-11-29 20:09:54,172 - INFO - allennlp.common.params - dataset_reader.token_indexers.roberta.model_name = /home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688 2022-11-29 20:09:54,172 - INFO - allennlp.common.params - dataset_reader.token_indexers.roberta.namespace = tags 2022-11-29 20:09:54,172 - INFO - allennlp.common.params - dataset_reader.token_indexers.roberta.max_length = None 2022-11-29 20:09:54,172 - INFO - transformers.configuration_utils - loading configuration file /home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688/config.json 2022-11-29 20:09:54,173 - INFO - transformers.configuration_utils - Model config RobertaConfig { "architectures": [ "RobertaModel" ], "attention_probs_dropout_prob": 0.1, "bos_token_id": 0, "do_sample": false, "eos_token_id": 2, "eos_token_ids": 0, "finetuning_task": null, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "hidden_size": 768, "id2label": { "0": "LABEL_0", "1": "LABEL_1" }, "initializer_range": 0.02, "intermediate_size": 3072, "is_decoder": false, "label2id": { "LABEL_0": 0, "LABEL_1": 1 }, "layer_norm_eps": 1e-05, "length_penalty": 1.0, "max_length": 20, "max_position_embeddings": 514, "model_type": "roberta", "num_attention_heads": 12, "num_beams": 1, "num_hidden_layers": 12, "num_labels": 2, "num_return_sequences": 1, "output_attentions": false, "output_hidden_states": false, "output_past": true, "pad_token_id": 1, "pruned_heads": {}, "repetition_penalty": 1.0, "temperature": 1.0, "top_k": 50, "top_p": 1.0, "torchscript": false, "type_vocab_size": 1, "use_bfloat16": false, "vocab_size": 50265 }
2022-11-29 20:09:54,173 - INFO - transformers.tokenization_utils - Model name '/home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688' not found in model shortcut name list (roberta-base, roberta-large, roberta-large-mnli, distilroberta-base, roberta-base-openai-detector, roberta-large-openai-detector). Assuming '/home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688' is a path, a model identifier, or url to a directory containing tokenizer files. 2022-11-29 20:09:54,173 - INFO - transformers.tokenization_utils - Didn't find file /home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688/added_tokens.json. We won't load it. 2022-11-29 20:09:54,173 - INFO - transformers.tokenization_utils - loading file /home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688/vocab.json 2022-11-29 20:09:54,173 - INFO - transformers.tokenization_utils - loading file /home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688/merges.txt 2022-11-29 20:09:54,173 - INFO - transformers.tokenization_utils - loading file None 2022-11-29 20:09:54,173 - INFO - transformers.tokenization_utils - loading file /home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688/special_tokens_map.json 2022-11-29 20:09:54,173 - INFO - transformers.tokenization_utils - loading file /home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688/tokenizer_config.json 2022-11-29 20:09:54,217 - INFO - allennlp.common.from_params - instantiating class <class 'allennlp.data.tokenizers.tokenizer.Tokenizer'> from params {'model_name': '/home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688', 'type': 'pretrained_transformer'} and extras {'local_rank', 'serialization_dir', 'batch_weight_key'} 2022-11-29 20:09:54,217 - INFO - allennlp.common.params - dataset_reader.tokenizer.type = pretrained_transformer 2022-11-29 20:09:54,217 - INFO - allennlp.common.from_params - instantiating class <class 'allennlp.data.tokenizers.pretrained_transformer_tokenizer.PretrainedTransformerTokenizer'> from params {'model_name': '/home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688'} and extras {'local_rank', 'serialization_dir', 'batch_weight_key'} 2022-11-29 20:09:54,217 - INFO - allennlp.common.params - dataset_reader.tokenizer.model_name = /home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688 2022-11-29 20:09:54,217 - INFO - allennlp.common.params - dataset_reader.tokenizer.add_special_tokens = True 2022-11-29 20:09:54,217 - INFO - allennlp.common.params - dataset_reader.tokenizer.max_length = None 2022-11-29 20:09:54,217 - INFO - allennlp.common.params - dataset_reader.tokenizer.stride = 0 2022-11-29 20:09:54,218 - INFO - allennlp.common.params - dataset_reader.tokenizer.truncation_strategy = longest_first 2022-11-29 20:09:54,218 - INFO - allennlp.common.params - dataset_reader.tokenizer.calculate_character_offsets = False 2022-11-29 20:09:54,218 - INFO - allennlp.common.params - dataset_reader.tokenizer.tokenizer_kwargs = None 2022-11-29 20:09:54,218 - INFO - transformers.configuration_utils - loading configuration file /home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688/config.json 2022-11-29 20:09:54,218 - INFO - transformers.configuration_utils - Model config RobertaConfig { "architectures": [ "RobertaModel" ], "attention_probs_dropout_prob": 0.1, "bos_token_id": 0, "do_sample": false, "eos_token_id": 2, "eos_token_ids": 0, "finetuning_task": null, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "hidden_size": 768, "id2label": { "0": "LABEL_0", "1": "LABEL_1" }, "initializer_range": 0.02, "intermediate_size": 3072, "is_decoder": false, "label2id": { "LABEL_0": 0, "LABEL_1": 1 }, "layer_norm_eps": 1e-05, "length_penalty": 1.0, "max_length": 20, "max_position_embeddings": 514, "model_type": "roberta", "num_attention_heads": 12, "num_beams": 1, "num_hidden_layers": 12, "num_labels": 2, "num_return_sequences": 1, "output_attentions": false, "output_hidden_states": false, "output_past": true, "pad_token_id": 1, "pruned_heads": {}, "repetition_penalty": 1.0, "temperature": 1.0, "top_k": 50, "top_p": 1.0, "torchscript": false, "type_vocab_size": 1, "use_bfloat16": false, "vocab_size": 50265 }
2022-11-29 20:09:54,218 - INFO - transformers.tokenization_utils - Model name '/home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688' not found in model shortcut name list (roberta-base, roberta-large, roberta-large-mnli, distilroberta-base, roberta-base-openai-detector, roberta-large-openai-detector). Assuming '/home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688' is a path, a model identifier, or url to a directory containing tokenizer files. 2022-11-29 20:09:54,218 - INFO - transformers.tokenization_utils - Didn't find file /home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688/added_tokens.json. We won't load it. 2022-11-29 20:09:54,218 - INFO - transformers.tokenization_utils - loading file /home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688/vocab.json 2022-11-29 20:09:54,218 - INFO - transformers.tokenization_utils - loading file /home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688/merges.txt 2022-11-29 20:09:54,218 - INFO - transformers.tokenization_utils - loading file None 2022-11-29 20:09:54,218 - INFO - transformers.tokenization_utils - loading file /home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688/special_tokens_map.json 2022-11-29 20:09:54,218 - INFO - transformers.tokenization_utils - loading file /home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688/tokenizer_config.json 2022-11-29 20:09:54,262 - INFO - allennlp.common.params - dataset_reader.max_sequence_length = 512 2022-11-29 20:09:54,262 - INFO - allennlp.common.params - dataset_reader.sample = None 2022-11-29 20:09:54,262 - INFO - allennlp.common.params - dataset_reader.skip_label_indexing = False 2022-11-29 20:09:54,262 - INFO - allennlp.common.params - dataset_reader.lazy = False 2022-11-29 20:09:54,263 - INFO - allennlp.common.params - train_data_path = https://s3-us-west-2.amazonaws.com/allennlp/dont_stop_pretraining/data/citation_intent/train.jsonl 2022-11-29 20:09:54,263 - INFO - allennlp.common.from_params - instantiating class <class 'allennlp.data.iterators.data_iterator.DataIterator'> from params {'batch_size': 16, 'sorting_keys': [['tokens', 'num_tokens']], 'type': 'bucket'} and extras {'local_rank', 'serialization_dir', 'batch_weight_key'} 2022-11-29 20:09:54,263 - INFO - allennlp.common.params - iterator.type = bucket 2022-11-29 20:09:54,263 - INFO - allennlp.common.from_params - instantiating class <class 'allennlp.data.iterators.bucket_iterator.BucketIterator'> from params {'batch_size': 16, 'sorting_keys': [['tokens', 'num_tokens']]} and extras {'local_rank', 'serialization_dir', 'batch_weight_key'} 2022-11-29 20:09:54,263 - INFO - allennlp.common.params - iterator.sorting_keys = [['tokens', 'num_tokens']] 2022-11-29 20:09:54,263 - INFO - allennlp.common.params - iterator.padding_noise = 0.1 2022-11-29 20:09:54,263 - INFO - allennlp.common.params - iterator.biggest_batch_first = False 2022-11-29 20:09:54,263 - INFO - allennlp.common.params - iterator.batch_size = 16 2022-11-29 20:09:54,263 - INFO - allennlp.common.params - iterator.instances_per_epoch = None 2022-11-29 20:09:54,263 - INFO - allennlp.common.params - iterator.max_instances_in_memory = None 2022-11-29 20:09:54,263 - INFO - allennlp.common.params - iterator.cache_instances = False 2022-11-29 20:09:54,263 - INFO - allennlp.common.params - iterator.track_epoch = False 2022-11-29 20:09:54,263 - INFO - allennlp.common.params - iterator.maximum_samples_per_batch = None 2022-11-29 20:09:54,263 - INFO - allennlp.common.params - iterator.skip_smaller_batches = False 2022-11-29 20:09:54,264 - INFO - allennlp.common.params - vocabulary = None 2022-11-29 20:09:54,264 - INFO - allennlp.common.params - datasets_for_vocab_creation = None 2022-11-29 20:09:54,264 - INFO - allennlp.common.from_params - instantiating class <class 'allennlp.data.dataset_readers.dataset_reader.DatasetReader'> from params {'lazy': False, 'max_sequence_length': 512, 'token_indexers': {'roberta': {'model_name': '/home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688', 'type': 'pretrained_transformer'}}, 'tokenizer': {'model_name': '/home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688', 'type': 'pretrained_transformer'}, 'type': 'text_classification_json_with_sampling'} and extras {'local_rank', 'serialization_dir', 'batch_weight_key'} 2022-11-29 20:09:54,264 - INFO - allennlp.common.params - validation_dataset_reader.type = text_classification_json_with_sampling 2022-11-29 20:09:54,264 - INFO - allennlp.common.from_params - instantiating class <class 'dont_stop_pretraining.data.dataset_readers.text_classification_json_reader_with_sampling.TextClassificationJsonReaderWithSampling'> from params {'lazy': False, 'max_sequence_length': 512, 'token_indexers': {'roberta': {'model_name': '/home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688', 'type': 'pretrained_transformer'}}, 'tokenizer': {'model_name': '/home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688', 'type': 'pretrained_transformer'}} and extras {'local_rank', 'serialization_dir', 'batch_weight_key'} 2022-11-29 20:09:54,264 - INFO - allennlp.common.from_params - instantiating class <class 'allennlp.data.token_indexers.token_indexer.TokenIndexer'> from params {'model_name': '/home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688', 'type': 'pretrained_transformer'} and extras {'local_rank', 'serialization_dir', 'batch_weight_key'} 2022-11-29 20:09:54,264 - INFO - allennlp.common.params - validation_dataset_reader.token_indexers.roberta.type = pretrained_transformer 2022-11-29 20:09:54,264 - INFO - allennlp.common.from_params - instantiating class <class 'allennlp.data.token_indexers.pretrained_transformer_indexer.PretrainedTransformerIndexer'> from params {'model_name': '/home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688'} and extras {'local_rank', 'serialization_dir', 'batch_weight_key'} 2022-11-29 20:09:54,264 - INFO - allennlp.common.params - validation_dataset_reader.token_indexers.roberta.token_min_padding_length = 0 2022-11-29 20:09:54,264 - INFO - allennlp.common.params - validation_dataset_reader.token_indexers.roberta.model_name = /home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688 2022-11-29 20:09:54,264 - INFO - allennlp.common.params - validation_dataset_reader.token_indexers.roberta.namespace = tags 2022-11-29 20:09:54,264 - INFO - allennlp.common.params - validation_dataset_reader.token_indexers.roberta.max_length = None 2022-11-29 20:09:54,265 - INFO - transformers.configuration_utils - loading configuration file /home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688/config.json 2022-11-29 20:09:54,265 - INFO - transformers.configuration_utils - Model config RobertaConfig { "architectures": [ "RobertaModel" ], "attention_probs_dropout_prob": 0.1, "bos_token_id": 0, "do_sample": false, "eos_token_id": 2, "eos_token_ids": 0, "finetuning_task": null, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "hidden_size": 768, "id2label": { "0": "LABEL_0", "1": "LABEL_1" }, "initializer_range": 0.02, "intermediate_size": 3072, "is_decoder": false, "label2id": { "LABEL_0": 0, "LABEL_1": 1 }, "layer_norm_eps": 1e-05, "length_penalty": 1.0, "max_length": 20, "max_position_embeddings": 514, "model_type": "roberta", "num_attention_heads": 12, "num_beams": 1, "num_hidden_layers": 12, "num_labels": 2, "num_return_sequences": 1, "output_attentions": false, "output_hidden_states": false, "output_past": true, "pad_token_id": 1, "pruned_heads": {}, "repetition_penalty": 1.0, "temperature": 1.0, "top_k": 50, "top_p": 1.0, "torchscript": false, "type_vocab_size": 1, "use_bfloat16": false, "vocab_size": 50265 }
2022-11-29 20:09:54,265 - INFO - transformers.tokenization_utils - Model name '/home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688' not found in model shortcut name list (roberta-base, roberta-large, roberta-large-mnli, distilroberta-base, roberta-base-openai-detector, roberta-large-openai-detector). Assuming '/home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688' is a path, a model identifier, or url to a directory containing tokenizer files. 2022-11-29 20:09:54,265 - INFO - transformers.tokenization_utils - Didn't find file /home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688/added_tokens.json. We won't load it. 2022-11-29 20:09:54,265 - INFO - transformers.tokenization_utils - loading file /home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688/vocab.json 2022-11-29 20:09:54,265 - INFO - transformers.tokenization_utils - loading file /home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688/merges.txt 2022-11-29 20:09:54,265 - INFO - transformers.tokenization_utils - loading file None 2022-11-29 20:09:54,265 - INFO - transformers.tokenization_utils - loading file /home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688/special_tokens_map.json 2022-11-29 20:09:54,265 - INFO - transformers.tokenization_utils - loading file /home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688/tokenizer_config.json 2022-11-29 20:09:54,309 - INFO - allennlp.common.from_params - instantiating class <class 'allennlp.data.tokenizers.tokenizer.Tokenizer'> from params {'model_name': '/home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688', 'type': 'pretrained_transformer'} and extras {'local_rank', 'serialization_dir', 'batch_weight_key'} 2022-11-29 20:09:54,309 - INFO - allennlp.common.params - validation_dataset_reader.tokenizer.type = pretrained_transformer 2022-11-29 20:09:54,309 - INFO - allennlp.common.from_params - instantiating class <class 'allennlp.data.tokenizers.pretrained_transformer_tokenizer.PretrainedTransformerTokenizer'> from params {'model_name': '/home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688'} and extras {'local_rank', 'serialization_dir', 'batch_weight_key'} 2022-11-29 20:09:54,309 - INFO - allennlp.common.params - validation_dataset_reader.tokenizer.model_name = /home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688 2022-11-29 20:09:54,309 - INFO - allennlp.common.params - validation_dataset_reader.tokenizer.add_special_tokens = True 2022-11-29 20:09:54,309 - INFO - allennlp.common.params - validation_dataset_reader.tokenizer.max_length = None 2022-11-29 20:09:54,310 - INFO - allennlp.common.params - validation_dataset_reader.tokenizer.stride = 0 2022-11-29 20:09:54,310 - INFO - allennlp.common.params - validation_dataset_reader.tokenizer.truncation_strategy = longest_first 2022-11-29 20:09:54,310 - INFO - allennlp.common.params - validation_dataset_reader.tokenizer.calculate_character_offsets = False 2022-11-29 20:09:54,310 - INFO - allennlp.common.params - validation_dataset_reader.tokenizer.tokenizer_kwargs = None 2022-11-29 20:09:54,310 - INFO - transformers.configuration_utils - loading configuration file /home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688/config.json 2022-11-29 20:09:54,310 - INFO - transformers.configuration_utils - Model config RobertaConfig { "architectures": [ "RobertaModel" ], "attention_probs_dropout_prob": 0.1, "bos_token_id": 0, "do_sample": false, "eos_token_id": 2, "eos_token_ids": 0, "finetuning_task": null, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "hidden_size": 768, "id2label": { "0": "LABEL_0", "1": "LABEL_1" }, "initializer_range": 0.02, "intermediate_size": 3072, "is_decoder": false, "label2id": { "LABEL_0": 0, "LABEL_1": 1 }, "layer_norm_eps": 1e-05, "length_penalty": 1.0, "max_length": 20, "max_position_embeddings": 514, "model_type": "roberta", "num_attention_heads": 12, "num_beams": 1, "num_hidden_layers": 12, "num_labels": 2, "num_return_sequences": 1, "output_attentions": false, "output_hidden_states": false, "output_past": true, "pad_token_id": 1, "pruned_heads": {}, "repetition_penalty": 1.0, "temperature": 1.0, "top_k": 50, "top_p": 1.0, "torchscript": false, "type_vocab_size": 1, "use_bfloat16": false, "vocab_size": 50265 }
2022-11-29 20:09:54,310 - INFO - transformers.tokenization_utils - Model name '/home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688' not found in model shortcut name list (roberta-base, roberta-large, roberta-large-mnli, distilroberta-base, roberta-base-openai-detector, roberta-large-openai-detector). Assuming '/home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688' is a path, a model identifier, or url to a directory containing tokenizer files. 2022-11-29 20:09:54,310 - INFO - transformers.tokenization_utils - Didn't find file /home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688/added_tokens.json. We won't load it. 2022-11-29 20:09:54,310 - INFO - transformers.tokenization_utils - loading file /home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688/vocab.json 2022-11-29 20:09:54,310 - INFO - transformers.tokenization_utils - loading file /home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688/merges.txt 2022-11-29 20:09:54,310 - INFO - transformers.tokenization_utils - loading file None 2022-11-29 20:09:54,310 - INFO - transformers.tokenization_utils - loading file /home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688/special_tokens_map.json 2022-11-29 20:09:54,310 - INFO - transformers.tokenization_utils - loading file /home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688/tokenizer_config.json 2022-11-29 20:09:54,355 - INFO - allennlp.common.params - validation_dataset_reader.max_sequence_length = 512 2022-11-29 20:09:54,355 - INFO - allennlp.common.params - validation_dataset_reader.sample = None 2022-11-29 20:09:54,355 - INFO - allennlp.common.params - validation_dataset_reader.skip_label_indexing = False 2022-11-29 20:09:54,355 - INFO - allennlp.common.params - validation_dataset_reader.lazy = False 2022-11-29 20:09:54,355 - INFO - allennlp.common.params - validation_data_path = https://s3-us-west-2.amazonaws.com/allennlp/dont_stop_pretraining/data/citation_intent/dev.jsonl 2022-11-29 20:09:54,355 - INFO - allennlp.common.from_params - instantiating class <class 'allennlp.data.iterators.data_iterator.DataIterator'> from params {'batch_size': 64, 'sorting_keys': [['tokens', 'num_tokens']], 'type': 'bucket'} and extras {'local_rank', 'serialization_dir', 'batch_weight_key'} 2022-11-29 20:09:54,355 - INFO - allennlp.common.params - validation_iterator.type = bucket 2022-11-29 20:09:54,355 - INFO - allennlp.common.from_params - instantiating class <class 'allennlp.data.iterators.bucket_iterator.BucketIterator'> from params {'batch_size': 64, 'sorting_keys': [['tokens', 'num_tokens']]} and extras {'local_rank', 'serialization_dir', 'batch_weight_key'} 2022-11-29 20:09:54,356 - INFO - allennlp.common.params - validation_iterator.sorting_keys = [['tokens', 'num_tokens']] 2022-11-29 20:09:54,356 - INFO - allennlp.common.params - validation_iterator.padding_noise = 0.1 2022-11-29 20:09:54,356 - INFO - allennlp.common.params - validation_iterator.biggest_batch_first = False 2022-11-29 20:09:54,356 - INFO - allennlp.common.params - validation_iterator.batch_size = 64 2022-11-29 20:09:54,356 - INFO - allennlp.common.params - validation_iterator.instances_per_epoch = None 2022-11-29 20:09:54,356 - INFO - allennlp.common.params - validation_iterator.max_instances_in_memory = None 2022-11-29 20:09:54,356 - INFO - allennlp.common.params - validation_iterator.cache_instances = False 2022-11-29 20:09:54,356 - INFO - allennlp.common.params - validation_iterator.track_epoch = False 2022-11-29 20:09:54,356 - INFO - allennlp.common.params - validation_iterator.maximum_samples_per_batch = None 2022-11-29 20:09:54,356 - INFO - allennlp.common.params - validation_iterator.skip_smaller_batches = False 2022-11-29 20:09:54,356 - INFO - allennlp.common.params - test_data_path = https://s3-us-west-2.amazonaws.com/allennlp/dont_stop_pretraining/data/citation_intent/test.jsonl 2022-11-29 20:09:54,356 - INFO - allennlp.common.params - evaluate_on_test = True 2022-11-29 20:09:54,356 - INFO - allennlp.training.util - Reading training data from https://s3-us-west-2.amazonaws.com/allennlp/dont_stop_pretraining/data/citation_intent/train.jsonl 1688it [00:01, 1207.73it/s] 2022-11-29 20:09:55,754 - INFO - allennlp.training.util - Reading validation data from https://s3-us-west-2.amazonaws.com/allennlp/dont_stop_pretraining/data/citation_intent/dev.jsonl 114it [00:00, 145.96it/s] 0it [00:00, ?it/s]2022-11-29 20:09:56,535 - INFO - allennlp.training.util - Reading test data from https://s3-us-west-2.amazonaws.com/allennlp/dont_stop_pretraining/data/citation_intent/test.jsonl 2022-11-29 20:09:57,283 - INFO - allennlp.data.vocabulary - Fitting token dictionary from dataset. 2022-11-29 20:09:57,293 - INFO - allennlp.common.from_params - instantiating class <class 'allennlp.models.model.Model'> from params {'dropout': '0.1', 'feedforward_layer': {'activations': 'tanh', 'hidden_dims': 768, 'input_dim': 768, 'num_layers': 1}, 'seq2vec_encoder': {'embedding_dim': 768, 'type': 'cls_pooler_x'}, 'text_field_embedder': {'roberta': {'model_name': '/home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688', 'type': 'pretrained_transformer'}}, 'type': 'basic_classifier_with_f1'} and extras {'local_rank', 'vocab', 'serialization_dir', 'batch_weight_key'} 2022-11-29 20:09:57,293 - INFO - allennlp.common.params - model.type = basic_classifier_with_f1 2022-11-29 20:09:57,293 - INFO - allennlp.common.from_params - instantiating class <class 'dont_stop_pretraining.models.basic_classifier_with_f1.BasicClassifierWithF1'> from params {'dropout': '0.1', 'feedforward_layer': {'activations': 'tanh', 'hidden_dims': 768, 'input_dim': 768, 'num_layers': 1}, 'seq2vec_encoder': {'embedding_dim': 768, 'type': 'cls_pooler_x'}, 'text_field_embedder': {'roberta': {'model_name': '/home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688', 'type': 'pretrained_transformer'}}} and extras {'local_rank', 'vocab', 'serialization_dir', 'batch_weight_key'} 2022-11-29 20:09:57,293 - INFO - allennlp.common.from_params - instantiating class <class 'allennlp.modules.text_field_embedders.text_field_embedder.TextFieldEmbedder'> from params {'roberta': {'model_name': '/home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688', 'type': 'pretrained_transformer'}} and extras {'local_rank', 'vocab', 'serialization_dir', 'batch_weight_key'} 2022-11-29 20:09:57,294 - INFO - allennlp.common.params - model.text_field_embedder.type = basic 2022-11-29 20:09:57,294 - INFO - allennlp.common.from_params - instantiating class <class 'allennlp.modules.text_field_embedders.basic_text_field_embedder.BasicTextFieldEmbedder'> from params {'roberta': {'model_name': '/home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688', 'type': 'pretrained_transformer'}} and extras {'local_rank', 'vocab', 'serialization_dir', 'batch_weight_key'} 139it [00:00, 186.05it/s] 1941it [00:00, 192557.63it/s] Traceback (most recent call last): File "/home/mikeleatila/anaconda3/envs/domains/lib/python3.7/site-packages/allennlp/common/params.py", line 243, in pop value = self.params.pop(key) KeyError: 'token_embedders'
During handling of the above exception, another exception occurred:
Traceback (most recent call last): File "/home/mikeleatila/anaconda3/envs/domains/bin/allennlp", line 8, in
sys.exit(run())
File "/home/mikeleatila/anaconda3/envs/domains/lib/python3.7/site-packages/allennlp/run.py", line 18, in run
main(prog="allennlp")
File "/home/mikeleatila/anaconda3/envs/domains/lib/python3.7/site-packages/allennlp/commands/init.py", line 93, in main
args.func(args)
File "/home/mikeleatila/anaconda3/envs/domains/lib/python3.7/site-packages/allennlp/commands/train.py", line 144, in train_model_from_args
dry_run=args.dry_run,
File "/home/mikeleatila/anaconda3/envs/domains/lib/python3.7/site-packages/allennlp/commands/train.py", line 203, in train_model_from_file
dry_run=dry_run,
File "/home/mikeleatila/anaconda3/envs/domains/lib/python3.7/site-packages/allennlp/commands/train.py", line 266, in train_model
dry_run=dry_run,
File "/home/mikeleatila/anaconda3/envs/domains/lib/python3.7/site-packages/allennlp/commands/train.py", line 450, in _train_worker
batch_weight_key=batch_weight_key,
File "/home/mikeleatila/anaconda3/envs/domains/lib/python3.7/site-packages/allennlp/common/from_params.py", line 555, in from_params
extras,
File "/home/mikeleatila/anaconda3/envs/domains/lib/python3.7/site-packages/allennlp/common/from_params.py", line 585, in from_params
return constructor_to_call(kwargs) # type: ignore
File "/home/mikeleatila/anaconda3/envs/domains/lib/python3.7/site-packages/allennlp/commands/train.py", line 652, in from_partialobjects
model = model.construct(vocab=vocabulary_)
File "/home/mikeleatila/anaconda3/envs/domains/lib/python3.7/site-packages/allennlp/common/lazy.py", line 29, in construct
return self._constructor(kwargs)
File "/home/mikeleatila/anaconda3/envs/domains/lib/python3.7/site-packages/allennlp/common/from_params.py", line 466, in constructor
return value_cls.from_params(params=popped_params, subextras)
File "/home/mikeleatila/anaconda3/envs/domains/lib/python3.7/site-packages/allennlp/common/from_params.py", line 555, in from_params
extras,
File "/home/mikeleatila/anaconda3/envs/domains/lib/python3.7/site-packages/allennlp/common/from_params.py", line 583, in from_params
kwargs = create_kwargs(constructor_to_inspect, cls, params, extras)
File "/home/mikeleatila/anaconda3/envs/domains/lib/python3.7/site-packages/allennlp/common/from_params.py", line 188, in create_kwargs
cls.name, param_name, annotation, param.default, params, extras
File "/home/mikeleatila/anaconda3/envs/domains/lib/python3.7/site-packages/allennlp/common/from_params.py", line 294, in pop_and_construct_arg
return construct_arg(class_name, name, popped_params, annotation, default, extras)
File "/home/mikeleatila/anaconda3/envs/domains/lib/python3.7/site-packages/allennlp/common/from_params.py", line 329, in construct_arg
return annotation.from_params(params=popped_params, subextras)
File "/home/mikeleatila/anaconda3/envs/domains/lib/python3.7/site-packages/allennlp/common/from_params.py", line 555, in from_params
extras,
File "/home/mikeleatila/anaconda3/envs/domains/lib/python3.7/site-packages/allennlp/common/from_params.py", line 583, in from_params
kwargs = create_kwargs(constructor_to_inspect, cls, params, extras)
File "/home/mikeleatila/anaconda3/envs/domains/lib/python3.7/site-packages/allennlp/common/from_params.py", line 188, in create_kwargs
cls.name, param_name, annotation, param.default, params, extras
File "/home/mikeleatila/anaconda3/envs/domains/lib/python3.7/site-packages/allennlp/common/from_params.py", line 287, in pop_and_construct_arg
popped_params = params.pop(name, default) if default != _NO_DEFAULT else params.pop(name)
File "/home/mikeleatila/anaconda3/envs/domains/lib/python3.7/site-packages/allennlp/common/params.py", line 245, in pop
raise ConfigurationError(f'key "{key}" is required at location "{self.history}"')
allennlp.common.checks.ConfigurationError: key "token_embedders" is required at location "model.text_field_embedder."
Traceback (most recent call last):
File "/home/mikeleatila/dont_stop_pretraining_master/scripts/train.py", line 142, in
main()
File "/home/mikeleatila/dont_stop_pretraining_master/scripts/train.py", line 139, in main
subprocess.run(" ".join(allennlp_command), shell=True, check=True)
File "/home/mikeleatila/anaconda3/envs/domains/lib/python3.7/subprocess.py", line 512, in run
output=stdout, stderr=stderr)
subprocess.CalledProcessError: Command 'allennlp train --include-package dont_stop_pretraining training_config/classifier.jsonnet -s model_logs/citation-intent-dapt-dapt' returned non-zero exit status 1.