Open Bachstelze opened 1 year ago
@Bachstelze Yeah, Mr. Peter Szemraj trained using flan-t5-large model, size of the model is 3.13 GB as you evaluated sir. Any suggetion, how can this flan-t5-large model from Mr. Peter Szemraj can be improved further.
@akesh1235 you probably mean flan-t5-large-grammar-synthesis. This model is only trained on the JFLEG dataset. You can add more datasets for the english-centric Flan-T5:
I have fine tune a flan-t5 model on my own small dataset about 1500 rows. But i find it not works. I only want to correct the error in my own dataset. others just to return the original text. it is not run as i expect.
We evaluated your and various T5-models on syntax: Evaluating PaLM-FLAN-T5 and previous models on syntax
Gramformer is the the grammar-tuned dot:
The evaluation suggests using FLAN-T5-base