-
I'm interested in your paper -- 'Input Combination Strategies for Multi-Source Transformer Decoder', Would you mind telling me how can I reproduce this work. I want cite this paper. Thanks
-
**Is your feature request related to a problem? Please describe.**
I want to use spaced repetition to train my vocabulary in foreign languages I learn.
**Describe the solution you'd like**
It wo…
-
It looks like in the vocab, the preferred method of defining types of (in this case) relations is to create a subclass which then becomes part of the vocabulary. @elf-pavlik has examples: Membership …
-
I am not able to figure out how to set the interactive model config to load `checkpoint_path='gs://scenic-bucket/ul2/ul220b/checkpoint_2650000'` from [here](https://github.com/google-research/google-r…
-
Hi
I want to train a RNNLM. My vocabulary size is 48603 (cutoff=100) or 72294 (cutoff=50). Data is stored in CTF format which is suitable for sparse data. Training RNN with vocabulary size of 4860…
-
Hi,
Thanks for providing the code. :)
I have a question regarding training the classifiers. What do you mean by replacing GPT2-large embeddings with roberta-base? I'm not sure if I totally und…
-
I'd like to see a feature that lets the user switch out main display of a word with an alternative writing. Or at least a way to disambiguate a word that has several alternate writings and meanings, b…
K-410 updated
9 years ago
-
When I run:
thumt-trainer \
--input corpus.tc.32k.zh.shuf corpus.tc.32k.en.shuf \
--vocabulary vocab.32k.zh.txt vocab.32k.en.txt \
--model transformer \
--validation newsdev2017.tc.32k.zh…
-
transformers version: 4.41.2
Python version: 3.12
System: windows
peft_config = LoraConfig(
task_type=TaskType.CAUSAL_LM,
target_modules=["q_proj", "k_proj", "v_…
-
### Bug description
I was able to fine-tune a 8B LLM using Huggingface training framework with PEFT+DeepSpeed stage 2 under fp16 precision(mixed precision training). Recently I would like to change…