Closed ameet-1997 closed 4 years ago
Hi Ameet,
You can find more details here: https://indicnlp.ai4bharat.org/
Specifically about the datasets here: https://indicnlp.ai4bharat.org/indic-glue/
And paper about this work here: https://indicnlp.ai4bharat.org/papers/arxiv2020_indicnlp_corpus.pdf
Thanks for the reply! I've gone through all these resources.
However, I was referring to documentation on the code for fine-tuning external models. Is this notebook (link) the only available resource for that? Or is there a place I can look for the documentation?
Hey Ameet. Unfortunately, we don't have a documentation yet. We are working on adding both user and developer documentation in our repository.
For now, depending on your use case, you can check if the following helps:
python3 -m fine_tune.cli --help
ALL_DATASETS
dictionary in https://github.com/AI4Bharat/indic-bert/blob/master/fine_tune/cli.py and use the corresponding key in the colab notebook. bert-base-multilingual-cased
) For list of all available names, please refer https://huggingface.co/transformers/pretrained_models.htmlIt is also possible to run the evaluation on non-huggingface models. I will shortly put a documentation on it and update here.
Thanks so much! Leaving this open so that you can close it after you have updated the documentation, but feel free to close it!
Hey, I have added some usage notes here https://github.com/AI4Bharat/indic-bert/blob/master/docs/advanced-usage.md. Does it answer your query?
Also, feel free to add here anything else that you'd want to see in the documentation.
Hi,
This repository has some great resources!
I was wondering where the documentation for the fine-tuning datasets are. Sorry if I overlooked something obvious. I want to evaluate my models (not IndicBERT) on your dataset, so wanted to see how that is possible.
Thanks!