-
ceifa updated
10 months ago
-
Thanks for the great work - I was wondering if you could upload the pretrained GPT2 model checkpoints (e.g. GPT2 Small & Medium C2S, NL+C2S) somewhere?
I am planning to use your work on several ot…
-
Hi @cg123,
Great library, thanks a lot, super useful!
I've finetuned GPT2 on 2 tasks (model1 and model2) and am trying to merge using your repo.
It turns out, using the original model as the…
-
Hi, I attempted to use speculative decoding but encountered some errors. May I ask for your assistance?
I used the parameters from the first example.
python ./examples/speculative_inference.py \…
-
After spending quite a bit of time and using a chunk of my resources code suddenly halted just to tell me that I need to "wait for 2.75s", but there is no option to continue research. The exception is…
-
论文introduction部分提到,“we calculate the similarity between the parameter-efficient tuning modules derived from GPT2-XL and GPT2-base or GPT2-medium”。
但是不同模型的LoRA module的tensor形状是不一样的,无法直接进行两两比较。请问这个相似度具…
-
I try to run the GPT2 training script. The version is 2.2.0. But there is no "lightseq.training.ops.pytorch.quantization" file. In which version, can I run the GPT2 training example?
-
I want to fine-tune the model and do some machine translation based on Gpt-2. I created my dataset according to the Gpt-2 paper in this format: 'sentence1 = translation1 \n sentence2 = translation2 \n…
-
Need to look into the docs for huggingface's pytorch-transformer library to see how to train it
then what to train it on? Gutenberg-dammit? that seems probably pretty good.
Or maybe just a subse…
-
Hi, i have just used the default params to p-tune the gpt2-medium on LAMA task and the results is as follows.
best dev_hit@1: 51.8 best test_hit@1: 44.5
For the results I got, I have some confusions…