dame-cell / Gaja

Other
2 stars 0 forks source link

Gaja

We release Gaja , a Hindi/Hinglish chat model instruction finetuned on SarvamAI's OpenHathi model.

Gajendra is a Hindi/Hinglish instruction-tuned model based on different instruct datasets.

This repository contains the code for "Gaja", a project focused on Fine-Tuning SarvamAI's OpenHathi model for Conversational task . which employs the LoRA + Unsloth methodology for efficient fine tuning.

Contents

1) Information 2) Indic-Eval 3) English-eval 4) Prompt-Format 5) Inference 6) Usage-Note

If you appreciate this work and found it helpful, consider giving it a star ⭐️ on GitHub. Your support motivates me to continue improving and adding new features. Thank you for your encouragement!

Information

Indic-Eval

Conducting a comprehensive zero-shot evaluation across various tasks, followed by the averaging of all scores, provides a holistic assessment of the model's performance.

Task # Samples Accuracy Precision F1 Recall BLEU Score Metrics
Indic-Sentiment Analysis 100 0.71 - 0.76 - - Accuracy, F1 score
Indic-QA Evaluation 50 - 0.62 0.68 0.75 - Bert Score
Indic-NLI 50 0.24 - 0.17 - - Accuracy, F1 score
Indic-Paraphrase 500 0.52 0.49 0.48 - Accuracy, F1 score, Precision

English-Eval

Model name Average ARC HellaSwag MMLU TruthfulQA Winogrande GSM8K
damerajee/Gaja-v1.00 47.69 52.82 76.31 40.83 44.64 70.64 0.91
manishiitg/open-aditi-hi-v2 59.31 59.39 82.01 61.41 45.84 77.19 30.02
ai4bharat/Airavata 45.52 46.5 69.26 43.9 40.62 68.82 4.02

Prompt-Format

The prompt for the Model without system prompt

<|im_start|>user
{}<|im_end|> 
<|im_start|>assistant
{}<|im_end|> 

The prompt for the Model with system prompt

|im_start|>system
{}<|im_end|> 
<|im_start|>user
{}<|im_end|> 
<|im_start|>assistant
{}<|im_end|> 

Inference

You can easily try chatting with this model on Huggingface spaces through this link Gaja

Note

Usage-Note

It's important to note that the models have not undergone detoxification. Therefore, while they possess impressive linguistic capabilities, there is a possibility for them to generate content that could be deemed harmful or offensive. We urge users to exercise discretion and supervise the model's outputs closely, especially in public or sensitive applications.