allenai / allennlp

An open-source NLP research library, built on PyTorch.
http://www.allennlp.org
Apache License 2.0
11.74k stars 2.25k forks source link

allennlp.common.checks.ConfigurationError: key "token_embedders" is required at location "model.text_field_embedder." #5736

Closed mikeleatila closed 1 year ago

mikeleatila commented 1 year ago

Hello, I am seeing this error - could you please help/advise how to fix it? Many thanks in advance!

/home/mikeleatila/anaconda3/envs/domains/bin/python /home/mikeleatila/dont_stop_pretraining_master/scripts/train.py --config training_config/classifier.jsonnet --serialization_dir model_logs/citation-intent-dapt-dapt --hyperparameters ROBERTA_CLASSIFIER_SMALL --dataset citation_intent --model /home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688 --device 0 --perf +f1 --evaluate_on_test 2022-11-29 20:09:53,526 - INFO - transformers.file_utils - PyTorch version 1.13.0 available. 2022-11-29 20:09:54,138 - INFO - pytorch_pretrained_bert.modeling - Better speed can be achieved with apex installed from https://www.github.com/nvidia/apex . 2022-11-29 20:09:54,170 - INFO - allennlp.common.params - random_seed = 237386 2022-11-29 20:09:54,170 - INFO - allennlp.common.params - numpy_seed = 237386 2022-11-29 20:09:54,170 - INFO - allennlp.common.params - pytorch_seed = 237386 2022-11-29 20:09:54,171 - INFO - allennlp.common.checks - Pytorch version: 1.13.0 2022-11-29 20:09:54,171 - INFO - allennlp.common.from_params - instantiating class <class 'allennlp.commands.train.TrainModel'> from params {'trainer': {'cuda_device': 0, 'gradient_accumulation_batch_size': 16, 'num_epochs': 10, 'num_serialized_models_to_keep': 0, 'optimizer': {'b1': 0.9, 'b2': 0.98, 'e': 1e-06, 'lr': '2e-05', 'max_grad_norm': 1, 'parameter_groups': [[['bias', 'LayerNorm.bias', 'LayerNorm.weight', 'layer_norm.weight'], {'weight_decay': 0}, []]], 'schedule': 'warmup_linear', 't_total': -1, 'type': 'bert_adam', 'warmup': 0.06, 'weight_decay': 0.1}, 'patience': 3, 'validation_metric': '+f1'}, 'model': {'dropout': '0.1', 'feedforward_layer': {'activations': 'tanh', 'hidden_dims': 768, 'input_dim': 768, 'num_layers': 1}, 'seq2vec_encoder': {'embedding_dim': 768, 'type': 'cls_pooler_x'}, 'text_field_embedder': {'roberta': {'model_name': '/home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688', 'type': 'pretrained_transformer'}}, 'type': 'basic_classifier_with_f1'}, 'dataset_reader': {'lazy': False, 'max_sequence_length': 512, 'token_indexers': {'roberta': {'model_name': '/home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688', 'type': 'pretrained_transformer'}}, 'tokenizer': {'model_name': '/home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688', 'type': 'pretrained_transformer'}, 'type': 'text_classification_json_with_sampling'}, 'validation_data_path': 'https://s3-us-west-2.amazonaws.com/allennlp/dont_stop_pretraining/data/citation_intent/dev.jsonl', 'evaluate_on_test': True, 'iterator': {'batch_size': 16, 'sorting_keys': [['tokens', 'num_tokens']], 'type': 'bucket'}, 'test_data_path': 'https://s3-us-west-2.amazonaws.com/allennlp/dont_stop_pretraining/data/citation_intent/test.jsonl', 'validation_iterator': {'batch_size': 64, 'sorting_keys': [['tokens', 'num_tokens']], 'type': 'bucket'}, 'validation_dataset_reader': {'lazy': False, 'max_sequence_length': 512, 'token_indexers': {'roberta': {'model_name': '/home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688', 'type': 'pretrained_transformer'}}, 'tokenizer': {'model_name': '/home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688', 'type': 'pretrained_transformer'}, 'type': 'text_classification_json_with_sampling'}, 'train_data_path': 'https://s3-us-west-2.amazonaws.com/allennlp/dont_stop_pretraining/data/citation_intent/train.jsonl'} and extras {'local_rank', 'serialization_dir', 'batch_weight_key'} 2022-11-29 20:09:54,171 - INFO - allennlp.common.params - type = default 2022-11-29 20:09:54,171 - INFO - allennlp.common.from_params - instantiating class <class 'allennlp.commands.train.TrainModel'> from params {'trainer': {'cuda_device': 0, 'gradient_accumulation_batch_size': 16, 'num_epochs': 10, 'num_serialized_models_to_keep': 0, 'optimizer': {'b1': 0.9, 'b2': 0.98, 'e': 1e-06, 'lr': '2e-05', 'max_grad_norm': 1, 'parameter_groups': [[['bias', 'LayerNorm.bias', 'LayerNorm.weight', 'layer_norm.weight'], {'weight_decay': 0}, []]], 'schedule': 'warmup_linear', 't_total': -1, 'type': 'bert_adam', 'warmup': 0.06, 'weight_decay': 0.1}, 'patience': 3, 'validation_metric': '+f1'}, 'model': {'dropout': '0.1', 'feedforward_layer': {'activations': 'tanh', 'hidden_dims': 768, 'input_dim': 768, 'num_layers': 1}, 'seq2vec_encoder': {'embedding_dim': 768, 'type': 'cls_pooler_x'}, 'text_field_embedder': {'roberta': {'model_name': '/home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688', 'type': 'pretrained_transformer'}}, 'type': 'basic_classifier_with_f1'}, 'dataset_reader': {'lazy': False, 'max_sequence_length': 512, 'token_indexers': {'roberta': {'model_name': '/home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688', 'type': 'pretrained_transformer'}}, 'tokenizer': {'model_name': '/home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688', 'type': 'pretrained_transformer'}, 'type': 'text_classification_json_with_sampling'}, 'validation_data_path': 'https://s3-us-west-2.amazonaws.com/allennlp/dont_stop_pretraining/data/citation_intent/dev.jsonl', 'evaluate_on_test': True, 'iterator': {'batch_size': 16, 'sorting_keys': [['tokens', 'num_tokens']], 'type': 'bucket'}, 'test_data_path': 'https://s3-us-west-2.amazonaws.com/allennlp/dont_stop_pretraining/data/citation_intent/test.jsonl', 'validation_iterator': {'batch_size': 64, 'sorting_keys': [['tokens', 'num_tokens']], 'type': 'bucket'}, 'validation_dataset_reader': {'lazy': False, 'max_sequence_length': 512, 'token_indexers': {'roberta': {'model_name': '/home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688', 'type': 'pretrained_transformer'}}, 'tokenizer': {'model_name': '/home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688', 'type': 'pretrained_transformer'}, 'type': 'text_classification_json_with_sampling'}, 'train_data_path': 'https://s3-us-west-2.amazonaws.com/allennlp/dont_stop_pretraining/data/citation_intent/train.jsonl'} and extras {'local_rank', 'serialization_dir', 'batch_weight_key'} 2022-11-29 20:09:54,171 - INFO - allennlp.common.from_params - instantiating class <class 'allennlp.data.dataset_readers.dataset_reader.DatasetReader'> from params {'lazy': False, 'max_sequence_length': 512, 'token_indexers': {'roberta': {'model_name': '/home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688', 'type': 'pretrained_transformer'}}, 'tokenizer': {'model_name': '/home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688', 'type': 'pretrained_transformer'}, 'type': 'text_classification_json_with_sampling'} and extras {'local_rank', 'serialization_dir', 'batch_weight_key'} 2022-11-29 20:09:54,172 - INFO - allennlp.common.params - dataset_reader.type = text_classification_json_with_sampling 2022-11-29 20:09:54,172 - INFO - allennlp.common.from_params - instantiating class <class 'dont_stop_pretraining.data.dataset_readers.text_classification_json_reader_with_sampling.TextClassificationJsonReaderWithSampling'> from params {'lazy': False, 'max_sequence_length': 512, 'token_indexers': {'roberta': {'model_name': '/home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688', 'type': 'pretrained_transformer'}}, 'tokenizer': {'model_name': '/home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688', 'type': 'pretrained_transformer'}} and extras {'local_rank', 'serialization_dir', 'batch_weight_key'} 2022-11-29 20:09:54,172 - INFO - allennlp.common.from_params - instantiating class <class 'allennlp.data.token_indexers.token_indexer.TokenIndexer'> from params {'model_name': '/home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688', 'type': 'pretrained_transformer'} and extras {'local_rank', 'serialization_dir', 'batch_weight_key'} 2022-11-29 20:09:54,172 - INFO - allennlp.common.params - dataset_reader.token_indexers.roberta.type = pretrained_transformer 2022-11-29 20:09:54,172 - INFO - allennlp.common.from_params - instantiating class <class 'allennlp.data.token_indexers.pretrained_transformer_indexer.PretrainedTransformerIndexer'> from params {'model_name': '/home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688'} and extras {'local_rank', 'serialization_dir', 'batch_weight_key'} 2022-11-29 20:09:54,172 - INFO - allennlp.common.params - dataset_reader.token_indexers.roberta.token_min_padding_length = 0 2022-11-29 20:09:54,172 - INFO - allennlp.common.params - dataset_reader.token_indexers.roberta.model_name = /home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688 2022-11-29 20:09:54,172 - INFO - allennlp.common.params - dataset_reader.token_indexers.roberta.namespace = tags 2022-11-29 20:09:54,172 - INFO - allennlp.common.params - dataset_reader.token_indexers.roberta.max_length = None 2022-11-29 20:09:54,172 - INFO - transformers.configuration_utils - loading configuration file /home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688/config.json 2022-11-29 20:09:54,173 - INFO - transformers.configuration_utils - Model config RobertaConfig { "architectures": [ "RobertaModel" ], "attention_probs_dropout_prob": 0.1, "bos_token_id": 0, "do_sample": false, "eos_token_id": 2, "eos_token_ids": 0, "finetuning_task": null, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "hidden_size": 768, "id2label": { "0": "LABEL_0", "1": "LABEL_1" }, "initializer_range": 0.02, "intermediate_size": 3072, "is_decoder": false, "label2id": { "LABEL_0": 0, "LABEL_1": 1 }, "layer_norm_eps": 1e-05, "length_penalty": 1.0, "max_length": 20, "max_position_embeddings": 514, "model_type": "roberta", "num_attention_heads": 12, "num_beams": 1, "num_hidden_layers": 12, "num_labels": 2, "num_return_sequences": 1, "output_attentions": false, "output_hidden_states": false, "output_past": true, "pad_token_id": 1, "pruned_heads": {}, "repetition_penalty": 1.0, "temperature": 1.0, "top_k": 50, "top_p": 1.0, "torchscript": false, "type_vocab_size": 1, "use_bfloat16": false, "vocab_size": 50265 }

2022-11-29 20:09:54,173 - INFO - transformers.tokenization_utils - Model name '/home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688' not found in model shortcut name list (roberta-base, roberta-large, roberta-large-mnli, distilroberta-base, roberta-base-openai-detector, roberta-large-openai-detector). Assuming '/home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688' is a path, a model identifier, or url to a directory containing tokenizer files. 2022-11-29 20:09:54,173 - INFO - transformers.tokenization_utils - Didn't find file /home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688/added_tokens.json. We won't load it. 2022-11-29 20:09:54,173 - INFO - transformers.tokenization_utils - loading file /home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688/vocab.json 2022-11-29 20:09:54,173 - INFO - transformers.tokenization_utils - loading file /home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688/merges.txt 2022-11-29 20:09:54,173 - INFO - transformers.tokenization_utils - loading file None 2022-11-29 20:09:54,173 - INFO - transformers.tokenization_utils - loading file /home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688/special_tokens_map.json 2022-11-29 20:09:54,173 - INFO - transformers.tokenization_utils - loading file /home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688/tokenizer_config.json 2022-11-29 20:09:54,217 - INFO - allennlp.common.from_params - instantiating class <class 'allennlp.data.tokenizers.tokenizer.Tokenizer'> from params {'model_name': '/home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688', 'type': 'pretrained_transformer'} and extras {'local_rank', 'serialization_dir', 'batch_weight_key'} 2022-11-29 20:09:54,217 - INFO - allennlp.common.params - dataset_reader.tokenizer.type = pretrained_transformer 2022-11-29 20:09:54,217 - INFO - allennlp.common.from_params - instantiating class <class 'allennlp.data.tokenizers.pretrained_transformer_tokenizer.PretrainedTransformerTokenizer'> from params {'model_name': '/home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688'} and extras {'local_rank', 'serialization_dir', 'batch_weight_key'} 2022-11-29 20:09:54,217 - INFO - allennlp.common.params - dataset_reader.tokenizer.model_name = /home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688 2022-11-29 20:09:54,217 - INFO - allennlp.common.params - dataset_reader.tokenizer.add_special_tokens = True 2022-11-29 20:09:54,217 - INFO - allennlp.common.params - dataset_reader.tokenizer.max_length = None 2022-11-29 20:09:54,217 - INFO - allennlp.common.params - dataset_reader.tokenizer.stride = 0 2022-11-29 20:09:54,218 - INFO - allennlp.common.params - dataset_reader.tokenizer.truncation_strategy = longest_first 2022-11-29 20:09:54,218 - INFO - allennlp.common.params - dataset_reader.tokenizer.calculate_character_offsets = False 2022-11-29 20:09:54,218 - INFO - allennlp.common.params - dataset_reader.tokenizer.tokenizer_kwargs = None 2022-11-29 20:09:54,218 - INFO - transformers.configuration_utils - loading configuration file /home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688/config.json 2022-11-29 20:09:54,218 - INFO - transformers.configuration_utils - Model config RobertaConfig { "architectures": [ "RobertaModel" ], "attention_probs_dropout_prob": 0.1, "bos_token_id": 0, "do_sample": false, "eos_token_id": 2, "eos_token_ids": 0, "finetuning_task": null, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "hidden_size": 768, "id2label": { "0": "LABEL_0", "1": "LABEL_1" }, "initializer_range": 0.02, "intermediate_size": 3072, "is_decoder": false, "label2id": { "LABEL_0": 0, "LABEL_1": 1 }, "layer_norm_eps": 1e-05, "length_penalty": 1.0, "max_length": 20, "max_position_embeddings": 514, "model_type": "roberta", "num_attention_heads": 12, "num_beams": 1, "num_hidden_layers": 12, "num_labels": 2, "num_return_sequences": 1, "output_attentions": false, "output_hidden_states": false, "output_past": true, "pad_token_id": 1, "pruned_heads": {}, "repetition_penalty": 1.0, "temperature": 1.0, "top_k": 50, "top_p": 1.0, "torchscript": false, "type_vocab_size": 1, "use_bfloat16": false, "vocab_size": 50265 }

2022-11-29 20:09:54,218 - INFO - transformers.tokenization_utils - Model name '/home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688' not found in model shortcut name list (roberta-base, roberta-large, roberta-large-mnli, distilroberta-base, roberta-base-openai-detector, roberta-large-openai-detector). Assuming '/home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688' is a path, a model identifier, or url to a directory containing tokenizer files. 2022-11-29 20:09:54,218 - INFO - transformers.tokenization_utils - Didn't find file /home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688/added_tokens.json. We won't load it. 2022-11-29 20:09:54,218 - INFO - transformers.tokenization_utils - loading file /home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688/vocab.json 2022-11-29 20:09:54,218 - INFO - transformers.tokenization_utils - loading file /home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688/merges.txt 2022-11-29 20:09:54,218 - INFO - transformers.tokenization_utils - loading file None 2022-11-29 20:09:54,218 - INFO - transformers.tokenization_utils - loading file /home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688/special_tokens_map.json 2022-11-29 20:09:54,218 - INFO - transformers.tokenization_utils - loading file /home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688/tokenizer_config.json 2022-11-29 20:09:54,262 - INFO - allennlp.common.params - dataset_reader.max_sequence_length = 512 2022-11-29 20:09:54,262 - INFO - allennlp.common.params - dataset_reader.sample = None 2022-11-29 20:09:54,262 - INFO - allennlp.common.params - dataset_reader.skip_label_indexing = False 2022-11-29 20:09:54,262 - INFO - allennlp.common.params - dataset_reader.lazy = False 2022-11-29 20:09:54,263 - INFO - allennlp.common.params - train_data_path = https://s3-us-west-2.amazonaws.com/allennlp/dont_stop_pretraining/data/citation_intent/train.jsonl 2022-11-29 20:09:54,263 - INFO - allennlp.common.from_params - instantiating class <class 'allennlp.data.iterators.data_iterator.DataIterator'> from params {'batch_size': 16, 'sorting_keys': [['tokens', 'num_tokens']], 'type': 'bucket'} and extras {'local_rank', 'serialization_dir', 'batch_weight_key'} 2022-11-29 20:09:54,263 - INFO - allennlp.common.params - iterator.type = bucket 2022-11-29 20:09:54,263 - INFO - allennlp.common.from_params - instantiating class <class 'allennlp.data.iterators.bucket_iterator.BucketIterator'> from params {'batch_size': 16, 'sorting_keys': [['tokens', 'num_tokens']]} and extras {'local_rank', 'serialization_dir', 'batch_weight_key'} 2022-11-29 20:09:54,263 - INFO - allennlp.common.params - iterator.sorting_keys = [['tokens', 'num_tokens']] 2022-11-29 20:09:54,263 - INFO - allennlp.common.params - iterator.padding_noise = 0.1 2022-11-29 20:09:54,263 - INFO - allennlp.common.params - iterator.biggest_batch_first = False 2022-11-29 20:09:54,263 - INFO - allennlp.common.params - iterator.batch_size = 16 2022-11-29 20:09:54,263 - INFO - allennlp.common.params - iterator.instances_per_epoch = None 2022-11-29 20:09:54,263 - INFO - allennlp.common.params - iterator.max_instances_in_memory = None 2022-11-29 20:09:54,263 - INFO - allennlp.common.params - iterator.cache_instances = False 2022-11-29 20:09:54,263 - INFO - allennlp.common.params - iterator.track_epoch = False 2022-11-29 20:09:54,263 - INFO - allennlp.common.params - iterator.maximum_samples_per_batch = None 2022-11-29 20:09:54,263 - INFO - allennlp.common.params - iterator.skip_smaller_batches = False 2022-11-29 20:09:54,264 - INFO - allennlp.common.params - vocabulary = None 2022-11-29 20:09:54,264 - INFO - allennlp.common.params - datasets_for_vocab_creation = None 2022-11-29 20:09:54,264 - INFO - allennlp.common.from_params - instantiating class <class 'allennlp.data.dataset_readers.dataset_reader.DatasetReader'> from params {'lazy': False, 'max_sequence_length': 512, 'token_indexers': {'roberta': {'model_name': '/home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688', 'type': 'pretrained_transformer'}}, 'tokenizer': {'model_name': '/home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688', 'type': 'pretrained_transformer'}, 'type': 'text_classification_json_with_sampling'} and extras {'local_rank', 'serialization_dir', 'batch_weight_key'} 2022-11-29 20:09:54,264 - INFO - allennlp.common.params - validation_dataset_reader.type = text_classification_json_with_sampling 2022-11-29 20:09:54,264 - INFO - allennlp.common.from_params - instantiating class <class 'dont_stop_pretraining.data.dataset_readers.text_classification_json_reader_with_sampling.TextClassificationJsonReaderWithSampling'> from params {'lazy': False, 'max_sequence_length': 512, 'token_indexers': {'roberta': {'model_name': '/home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688', 'type': 'pretrained_transformer'}}, 'tokenizer': {'model_name': '/home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688', 'type': 'pretrained_transformer'}} and extras {'local_rank', 'serialization_dir', 'batch_weight_key'} 2022-11-29 20:09:54,264 - INFO - allennlp.common.from_params - instantiating class <class 'allennlp.data.token_indexers.token_indexer.TokenIndexer'> from params {'model_name': '/home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688', 'type': 'pretrained_transformer'} and extras {'local_rank', 'serialization_dir', 'batch_weight_key'} 2022-11-29 20:09:54,264 - INFO - allennlp.common.params - validation_dataset_reader.token_indexers.roberta.type = pretrained_transformer 2022-11-29 20:09:54,264 - INFO - allennlp.common.from_params - instantiating class <class 'allennlp.data.token_indexers.pretrained_transformer_indexer.PretrainedTransformerIndexer'> from params {'model_name': '/home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688'} and extras {'local_rank', 'serialization_dir', 'batch_weight_key'} 2022-11-29 20:09:54,264 - INFO - allennlp.common.params - validation_dataset_reader.token_indexers.roberta.token_min_padding_length = 0 2022-11-29 20:09:54,264 - INFO - allennlp.common.params - validation_dataset_reader.token_indexers.roberta.model_name = /home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688 2022-11-29 20:09:54,264 - INFO - allennlp.common.params - validation_dataset_reader.token_indexers.roberta.namespace = tags 2022-11-29 20:09:54,264 - INFO - allennlp.common.params - validation_dataset_reader.token_indexers.roberta.max_length = None 2022-11-29 20:09:54,265 - INFO - transformers.configuration_utils - loading configuration file /home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688/config.json 2022-11-29 20:09:54,265 - INFO - transformers.configuration_utils - Model config RobertaConfig { "architectures": [ "RobertaModel" ], "attention_probs_dropout_prob": 0.1, "bos_token_id": 0, "do_sample": false, "eos_token_id": 2, "eos_token_ids": 0, "finetuning_task": null, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "hidden_size": 768, "id2label": { "0": "LABEL_0", "1": "LABEL_1" }, "initializer_range": 0.02, "intermediate_size": 3072, "is_decoder": false, "label2id": { "LABEL_0": 0, "LABEL_1": 1 }, "layer_norm_eps": 1e-05, "length_penalty": 1.0, "max_length": 20, "max_position_embeddings": 514, "model_type": "roberta", "num_attention_heads": 12, "num_beams": 1, "num_hidden_layers": 12, "num_labels": 2, "num_return_sequences": 1, "output_attentions": false, "output_hidden_states": false, "output_past": true, "pad_token_id": 1, "pruned_heads": {}, "repetition_penalty": 1.0, "temperature": 1.0, "top_k": 50, "top_p": 1.0, "torchscript": false, "type_vocab_size": 1, "use_bfloat16": false, "vocab_size": 50265 }

2022-11-29 20:09:54,265 - INFO - transformers.tokenization_utils - Model name '/home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688' not found in model shortcut name list (roberta-base, roberta-large, roberta-large-mnli, distilroberta-base, roberta-base-openai-detector, roberta-large-openai-detector). Assuming '/home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688' is a path, a model identifier, or url to a directory containing tokenizer files. 2022-11-29 20:09:54,265 - INFO - transformers.tokenization_utils - Didn't find file /home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688/added_tokens.json. We won't load it. 2022-11-29 20:09:54,265 - INFO - transformers.tokenization_utils - loading file /home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688/vocab.json 2022-11-29 20:09:54,265 - INFO - transformers.tokenization_utils - loading file /home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688/merges.txt 2022-11-29 20:09:54,265 - INFO - transformers.tokenization_utils - loading file None 2022-11-29 20:09:54,265 - INFO - transformers.tokenization_utils - loading file /home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688/special_tokens_map.json 2022-11-29 20:09:54,265 - INFO - transformers.tokenization_utils - loading file /home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688/tokenizer_config.json 2022-11-29 20:09:54,309 - INFO - allennlp.common.from_params - instantiating class <class 'allennlp.data.tokenizers.tokenizer.Tokenizer'> from params {'model_name': '/home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688', 'type': 'pretrained_transformer'} and extras {'local_rank', 'serialization_dir', 'batch_weight_key'} 2022-11-29 20:09:54,309 - INFO - allennlp.common.params - validation_dataset_reader.tokenizer.type = pretrained_transformer 2022-11-29 20:09:54,309 - INFO - allennlp.common.from_params - instantiating class <class 'allennlp.data.tokenizers.pretrained_transformer_tokenizer.PretrainedTransformerTokenizer'> from params {'model_name': '/home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688'} and extras {'local_rank', 'serialization_dir', 'batch_weight_key'} 2022-11-29 20:09:54,309 - INFO - allennlp.common.params - validation_dataset_reader.tokenizer.model_name = /home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688 2022-11-29 20:09:54,309 - INFO - allennlp.common.params - validation_dataset_reader.tokenizer.add_special_tokens = True 2022-11-29 20:09:54,309 - INFO - allennlp.common.params - validation_dataset_reader.tokenizer.max_length = None 2022-11-29 20:09:54,310 - INFO - allennlp.common.params - validation_dataset_reader.tokenizer.stride = 0 2022-11-29 20:09:54,310 - INFO - allennlp.common.params - validation_dataset_reader.tokenizer.truncation_strategy = longest_first 2022-11-29 20:09:54,310 - INFO - allennlp.common.params - validation_dataset_reader.tokenizer.calculate_character_offsets = False 2022-11-29 20:09:54,310 - INFO - allennlp.common.params - validation_dataset_reader.tokenizer.tokenizer_kwargs = None 2022-11-29 20:09:54,310 - INFO - transformers.configuration_utils - loading configuration file /home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688/config.json 2022-11-29 20:09:54,310 - INFO - transformers.configuration_utils - Model config RobertaConfig { "architectures": [ "RobertaModel" ], "attention_probs_dropout_prob": 0.1, "bos_token_id": 0, "do_sample": false, "eos_token_id": 2, "eos_token_ids": 0, "finetuning_task": null, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "hidden_size": 768, "id2label": { "0": "LABEL_0", "1": "LABEL_1" }, "initializer_range": 0.02, "intermediate_size": 3072, "is_decoder": false, "label2id": { "LABEL_0": 0, "LABEL_1": 1 }, "layer_norm_eps": 1e-05, "length_penalty": 1.0, "max_length": 20, "max_position_embeddings": 514, "model_type": "roberta", "num_attention_heads": 12, "num_beams": 1, "num_hidden_layers": 12, "num_labels": 2, "num_return_sequences": 1, "output_attentions": false, "output_hidden_states": false, "output_past": true, "pad_token_id": 1, "pruned_heads": {}, "repetition_penalty": 1.0, "temperature": 1.0, "top_k": 50, "top_p": 1.0, "torchscript": false, "type_vocab_size": 1, "use_bfloat16": false, "vocab_size": 50265 }

2022-11-29 20:09:54,310 - INFO - transformers.tokenization_utils - Model name '/home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688' not found in model shortcut name list (roberta-base, roberta-large, roberta-large-mnli, distilroberta-base, roberta-base-openai-detector, roberta-large-openai-detector). Assuming '/home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688' is a path, a model identifier, or url to a directory containing tokenizer files. 2022-11-29 20:09:54,310 - INFO - transformers.tokenization_utils - Didn't find file /home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688/added_tokens.json. We won't load it. 2022-11-29 20:09:54,310 - INFO - transformers.tokenization_utils - loading file /home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688/vocab.json 2022-11-29 20:09:54,310 - INFO - transformers.tokenization_utils - loading file /home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688/merges.txt 2022-11-29 20:09:54,310 - INFO - transformers.tokenization_utils - loading file None 2022-11-29 20:09:54,310 - INFO - transformers.tokenization_utils - loading file /home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688/special_tokens_map.json 2022-11-29 20:09:54,310 - INFO - transformers.tokenization_utils - loading file /home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688/tokenizer_config.json 2022-11-29 20:09:54,355 - INFO - allennlp.common.params - validation_dataset_reader.max_sequence_length = 512 2022-11-29 20:09:54,355 - INFO - allennlp.common.params - validation_dataset_reader.sample = None 2022-11-29 20:09:54,355 - INFO - allennlp.common.params - validation_dataset_reader.skip_label_indexing = False 2022-11-29 20:09:54,355 - INFO - allennlp.common.params - validation_dataset_reader.lazy = False 2022-11-29 20:09:54,355 - INFO - allennlp.common.params - validation_data_path = https://s3-us-west-2.amazonaws.com/allennlp/dont_stop_pretraining/data/citation_intent/dev.jsonl 2022-11-29 20:09:54,355 - INFO - allennlp.common.from_params - instantiating class <class 'allennlp.data.iterators.data_iterator.DataIterator'> from params {'batch_size': 64, 'sorting_keys': [['tokens', 'num_tokens']], 'type': 'bucket'} and extras {'local_rank', 'serialization_dir', 'batch_weight_key'} 2022-11-29 20:09:54,355 - INFO - allennlp.common.params - validation_iterator.type = bucket 2022-11-29 20:09:54,355 - INFO - allennlp.common.from_params - instantiating class <class 'allennlp.data.iterators.bucket_iterator.BucketIterator'> from params {'batch_size': 64, 'sorting_keys': [['tokens', 'num_tokens']]} and extras {'local_rank', 'serialization_dir', 'batch_weight_key'} 2022-11-29 20:09:54,356 - INFO - allennlp.common.params - validation_iterator.sorting_keys = [['tokens', 'num_tokens']] 2022-11-29 20:09:54,356 - INFO - allennlp.common.params - validation_iterator.padding_noise = 0.1 2022-11-29 20:09:54,356 - INFO - allennlp.common.params - validation_iterator.biggest_batch_first = False 2022-11-29 20:09:54,356 - INFO - allennlp.common.params - validation_iterator.batch_size = 64 2022-11-29 20:09:54,356 - INFO - allennlp.common.params - validation_iterator.instances_per_epoch = None 2022-11-29 20:09:54,356 - INFO - allennlp.common.params - validation_iterator.max_instances_in_memory = None 2022-11-29 20:09:54,356 - INFO - allennlp.common.params - validation_iterator.cache_instances = False 2022-11-29 20:09:54,356 - INFO - allennlp.common.params - validation_iterator.track_epoch = False 2022-11-29 20:09:54,356 - INFO - allennlp.common.params - validation_iterator.maximum_samples_per_batch = None 2022-11-29 20:09:54,356 - INFO - allennlp.common.params - validation_iterator.skip_smaller_batches = False 2022-11-29 20:09:54,356 - INFO - allennlp.common.params - test_data_path = https://s3-us-west-2.amazonaws.com/allennlp/dont_stop_pretraining/data/citation_intent/test.jsonl 2022-11-29 20:09:54,356 - INFO - allennlp.common.params - evaluate_on_test = True 2022-11-29 20:09:54,356 - INFO - allennlp.training.util - Reading training data from https://s3-us-west-2.amazonaws.com/allennlp/dont_stop_pretraining/data/citation_intent/train.jsonl 1688it [00:01, 1207.73it/s] 2022-11-29 20:09:55,754 - INFO - allennlp.training.util - Reading validation data from https://s3-us-west-2.amazonaws.com/allennlp/dont_stop_pretraining/data/citation_intent/dev.jsonl 114it [00:00, 145.96it/s] 0it [00:00, ?it/s]2022-11-29 20:09:56,535 - INFO - allennlp.training.util - Reading test data from https://s3-us-west-2.amazonaws.com/allennlp/dont_stop_pretraining/data/citation_intent/test.jsonl 2022-11-29 20:09:57,283 - INFO - allennlp.data.vocabulary - Fitting token dictionary from dataset. 2022-11-29 20:09:57,293 - INFO - allennlp.common.from_params - instantiating class <class 'allennlp.models.model.Model'> from params {'dropout': '0.1', 'feedforward_layer': {'activations': 'tanh', 'hidden_dims': 768, 'input_dim': 768, 'num_layers': 1}, 'seq2vec_encoder': {'embedding_dim': 768, 'type': 'cls_pooler_x'}, 'text_field_embedder': {'roberta': {'model_name': '/home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688', 'type': 'pretrained_transformer'}}, 'type': 'basic_classifier_with_f1'} and extras {'local_rank', 'vocab', 'serialization_dir', 'batch_weight_key'} 2022-11-29 20:09:57,293 - INFO - allennlp.common.params - model.type = basic_classifier_with_f1 2022-11-29 20:09:57,293 - INFO - allennlp.common.from_params - instantiating class <class 'dont_stop_pretraining.models.basic_classifier_with_f1.BasicClassifierWithF1'> from params {'dropout': '0.1', 'feedforward_layer': {'activations': 'tanh', 'hidden_dims': 768, 'input_dim': 768, 'num_layers': 1}, 'seq2vec_encoder': {'embedding_dim': 768, 'type': 'cls_pooler_x'}, 'text_field_embedder': {'roberta': {'model_name': '/home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688', 'type': 'pretrained_transformer'}}} and extras {'local_rank', 'vocab', 'serialization_dir', 'batch_weight_key'} 2022-11-29 20:09:57,293 - INFO - allennlp.common.from_params - instantiating class <class 'allennlp.modules.text_field_embedders.text_field_embedder.TextFieldEmbedder'> from params {'roberta': {'model_name': '/home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688', 'type': 'pretrained_transformer'}} and extras {'local_rank', 'vocab', 'serialization_dir', 'batch_weight_key'} 2022-11-29 20:09:57,294 - INFO - allennlp.common.params - model.text_field_embedder.type = basic 2022-11-29 20:09:57,294 - INFO - allennlp.common.from_params - instantiating class <class 'allennlp.modules.text_field_embedders.basic_text_field_embedder.BasicTextFieldEmbedder'> from params {'roberta': {'model_name': '/home/mikeleatila/dont_stop_pretraining_master/pretrained_models/dsp_roberta_base_dapt_cs_tapt_citation_intent_1688', 'type': 'pretrained_transformer'}} and extras {'local_rank', 'vocab', 'serialization_dir', 'batch_weight_key'} 139it [00:00, 186.05it/s] 1941it [00:00, 192557.63it/s] Traceback (most recent call last): File "/home/mikeleatila/anaconda3/envs/domains/lib/python3.7/site-packages/allennlp/common/params.py", line 243, in pop value = self.params.pop(key) KeyError: 'token_embedders'

During handling of the above exception, another exception occurred:

Traceback (most recent call last): File "/home/mikeleatila/anaconda3/envs/domains/bin/allennlp", line 8, in sys.exit(run()) File "/home/mikeleatila/anaconda3/envs/domains/lib/python3.7/site-packages/allennlp/run.py", line 18, in run main(prog="allennlp") File "/home/mikeleatila/anaconda3/envs/domains/lib/python3.7/site-packages/allennlp/commands/init.py", line 93, in main args.func(args) File "/home/mikeleatila/anaconda3/envs/domains/lib/python3.7/site-packages/allennlp/commands/train.py", line 144, in train_model_from_args dry_run=args.dry_run, File "/home/mikeleatila/anaconda3/envs/domains/lib/python3.7/site-packages/allennlp/commands/train.py", line 203, in train_model_from_file dry_run=dry_run, File "/home/mikeleatila/anaconda3/envs/domains/lib/python3.7/site-packages/allennlp/commands/train.py", line 266, in train_model dry_run=dry_run, File "/home/mikeleatila/anaconda3/envs/domains/lib/python3.7/site-packages/allennlp/commands/train.py", line 450, in _train_worker batch_weight_key=batch_weight_key, File "/home/mikeleatila/anaconda3/envs/domains/lib/python3.7/site-packages/allennlp/common/from_params.py", line 555, in from_params extras, File "/home/mikeleatila/anaconda3/envs/domains/lib/python3.7/site-packages/allennlp/common/from_params.py", line 585, in from_params return constructor_to_call(kwargs) # type: ignore File "/home/mikeleatila/anaconda3/envs/domains/lib/python3.7/site-packages/allennlp/commands/train.py", line 652, in from_partialobjects model = model.construct(vocab=vocabulary_) File "/home/mikeleatila/anaconda3/envs/domains/lib/python3.7/site-packages/allennlp/common/lazy.py", line 29, in construct return self._constructor(kwargs) File "/home/mikeleatila/anaconda3/envs/domains/lib/python3.7/site-packages/allennlp/common/from_params.py", line 466, in constructor return value_cls.from_params(params=popped_params, subextras) File "/home/mikeleatila/anaconda3/envs/domains/lib/python3.7/site-packages/allennlp/common/from_params.py", line 555, in from_params extras, File "/home/mikeleatila/anaconda3/envs/domains/lib/python3.7/site-packages/allennlp/common/from_params.py", line 583, in from_params kwargs = create_kwargs(constructor_to_inspect, cls, params, extras) File "/home/mikeleatila/anaconda3/envs/domains/lib/python3.7/site-packages/allennlp/common/from_params.py", line 188, in create_kwargs cls.name, param_name, annotation, param.default, params, extras File "/home/mikeleatila/anaconda3/envs/domains/lib/python3.7/site-packages/allennlp/common/from_params.py", line 294, in pop_and_construct_arg return construct_arg(class_name, name, popped_params, annotation, default, extras) File "/home/mikeleatila/anaconda3/envs/domains/lib/python3.7/site-packages/allennlp/common/from_params.py", line 329, in construct_arg return annotation.from_params(params=popped_params, subextras) File "/home/mikeleatila/anaconda3/envs/domains/lib/python3.7/site-packages/allennlp/common/from_params.py", line 555, in from_params extras, File "/home/mikeleatila/anaconda3/envs/domains/lib/python3.7/site-packages/allennlp/common/from_params.py", line 583, in from_params kwargs = create_kwargs(constructor_to_inspect, cls, params, extras) File "/home/mikeleatila/anaconda3/envs/domains/lib/python3.7/site-packages/allennlp/common/from_params.py", line 188, in create_kwargs cls.name, param_name, annotation, param.default, params, extras File "/home/mikeleatila/anaconda3/envs/domains/lib/python3.7/site-packages/allennlp/common/from_params.py", line 287, in pop_and_construct_arg popped_params = params.pop(name, default) if default != _NO_DEFAULT else params.pop(name) File "/home/mikeleatila/anaconda3/envs/domains/lib/python3.7/site-packages/allennlp/common/params.py", line 245, in pop raise ConfigurationError(f'key "{key}" is required at location "{self.history}"') allennlp.common.checks.ConfigurationError: key "token_embedders" is required at location "model.text_field_embedder." Traceback (most recent call last): File "/home/mikeleatila/dont_stop_pretraining_master/scripts/train.py", line 142, in main() File "/home/mikeleatila/dont_stop_pretraining_master/scripts/train.py", line 139, in main subprocess.run(" ".join(allennlp_command), shell=True, check=True) File "/home/mikeleatila/anaconda3/envs/domains/lib/python3.7/subprocess.py", line 512, in run output=stdout, stderr=stderr) subprocess.CalledProcessError: Command 'allennlp train --include-package dont_stop_pretraining training_config/classifier.jsonnet -s model_logs/citation-intent-dapt-dapt' returned non-zero exit status 1.

mikeleatila commented 1 year ago

@oroszgy @mfa @justindujardin @luckytoilet

snat1505027 commented 1 year ago

@mikeleatila, were you able to solve this? I am getting this error now!

github-actions[bot] commented 1 year ago

This issue is being closed due to lack of activity. If you think it still needs to be addressed, please comment on this thread 👇