Dijital-Twin / model

0 stars 0 forks source link

feat: Finetune DialoGPT Medium & Large #9

Closed t4r7k closed 6 months ago

t4r7k commented 6 months ago

Model Overview

DialoGPT is a SOTA large-scale pretrained dialogue response generation model for multiturn conversations. The human evaluation results indicate that the response generated from DialoGPT is comparable to human response quality under a single-turn conversation Turing test. The model is trained on 147M multi-turn dialogue from Reddit discussion thread.

Model URL

https://huggingface.co/docs/transformers/model_doc/dialogpt

t4r7k commented 6 months ago

DialoGPT Fine-tune Report

Commit URL

Introduction

This project involved exploring and fine-tuning language models available on Huggingface for generating responses that capture the essence of Rachel Green from the TV series "Friends". The DialoGPT medium and large models were chosen for this task due to their capabilities in generating human-like text.

Goal of This Issue

The aim was to fine-tune DialoGPT models to accurately reflect Rachel's personality in their responses, ensuring the outputs feel authentic and character-consistent.


Fine-tune DialoGPT Medium

Dialogues featuring Rachel from 227 episodes were used to fine-tune the DialoGPT medium model. The fine-tuning aimed at enhancing the model's ability to mimic Rachel's speaking style and character nuances.

Model Huggingface URL

microsoft/DialoGPT-medium

Training Parameters

Example Responses to Input Sentences


Fine-tune DialoGPT Large

The DialoGPT large model was also fine-tuned with the same dataset to compare its performance against the medium model in capturing Rachel's character.

Model Huggingface URL

microsoft/DialoGPT-large

Training Parameters

Example Responses to Input Sentences


Evaluation

Upon evaluating both versions of the fine-tuned DialoGPT models, it was observed that while both models effectively captured the essence of Rachel Green's character, the DialoGPT large model demonstrated a slightly more nuanced understanding of the character's personality and was able to generate responses with greater depth. Hovewer, although these models are good in single-turn conversations, demonstrating a nuanced understanding of Rachel Green's personality and generating responses with greater depth, they significantly underperform in multi-turn dialogues. Their inability to effectively refer back to previous context in these extended exchanges undermines their suitability as a chatbot for more dynamic interactions. That said, we decided to use other models for the project.