budzianowski / multiwoz

Source code for end-to-end dialogue model from the MultiWOZ paper (Budzianowski et al. 2018, EMNLP)
MIT License
866 stars 199 forks source link

Plan for including LLMs' Zero-shot performance? #132

Open Leezekun opened 4 months ago

Leezekun commented 4 months ago

Hi,

Thank you for the great work!

Given the current prevalence of Large Language Models (LLMs), are there any plans to include more LLM-based approaches in performance evaluations, especially focusing on zero-shot performance?

Here are a few relevant papers and approaches:

Are there any plans to benchmark the performance of LLMs in zero-shot settings? I would be happy to assist with this if needed.

budzianowski commented 4 months ago

Hi @Leezekun - thanks for posting this. A simple answer is - absolutely! There are numbers of efforts to work in a zero-shot manner. If you are happy to update the benchmarks that would be very helpful!