donttal / TARA

TARA
Apache License 2.0
4 stars 1 forks source link

TARA

The code of research paper Distinguishability Calibration to In-Context Learning.

This paper has been accepted at the The 17th Conference of the European Chapter of the Association for Computational Linguistics(EACL 2023 findings).

We currently provide a fully operational version of .ipynb.

Requirements

OpenPromptForTARA install

OpenPromptForTARA is a variant version for TARA based on OpenPrompt. Thanks for their development contribution.

cd OpenPromptForTARA
pip install -r requirements.txt
python setup.py install

Train

sh train.sh conf/your_train.conf

Configuration

# training_type
training_type = "mix"

# prompt_type
template_type = "ptuning"

# Data
dataset = "go_emotions"
data_condition = "fewshot"
num_examples_per_label = 50

# zero-shot
few_shot_train = True

# train
model = "BERT"
model_lr = 2e-5
template_lr = 1e-3
epoch_num = 3
batch_size = 8
use_cuda = True

# TML
bank_size = 32

Visualisation

cd data
wget https://nlp.stanford.edu/data/glove.6B.zip
unzip glove.6B.zip