mayubo2333 / LLM-IE

EMNLP'2023 (Findings): Large Language Model Is Not a Good Few-shot Information Extractor, but a Good Reranker for Hard Samples!
37 stars 3 forks source link

Large Language Model Is Not a Good Few-shot Information Extractor, but a Good Reranker for Hard Samples!

This is the implementation of filter-then-rerank pipeline in Large Language Model Is Not a Good Few-shot Information Extractor, but a Good Reranker for Hard Samples!. EMNLP'2023 (Findings).

If you have any question about this work, feel free to contact yubo001@e.ntu.edu.sg

Overview

We develop filter-then-rerank paradigm to solve IE tasks effectively and efficiently. Combining the merits of both Small Language Models (SLMs) and Large Language Models (LLMs), this paradigm includes two main components:

Update

Prerequisite

Environment

OpenAI API key

It you use GPT-4 as the reranker, please set up your openAI API key by:

export OPENAI_API_KEY=[your api key]

Results from the filter (SLM)

Unfortunately, we only provide the SLM's preliminary predictions of FewNERD dataset. We are not granted to release SLM's predictions on TACREV and ACE05 because they are non-public datasets. For reproduce the results of these two datasets, we suggest you:

  1. Run SLM-based methods (for example, vanilla fine-tuning) on these datasets.

  2. Save the results in ./SLM_results. For TACREV (relation extraction task) and ACE05 (event detection task), please save the predictions as the format shown in ./SLM_results/examples_re_tacrev.json and ./SLM_results/examples_ed_ace.json, respectively.

Quick start

FewNERD

MODEL_NAME=[gpt-4|Mistral-7B-Instruct-v0.2|...] bash ./scripts/run_ner_fewnerd.sh

TACREV

# To run this command, you shall firstly generate and save corresponding SLM's predictions.
# Then complete the env variables SLM_RES_PATH and OUT_PATH in the script.
MODEL_NAME=[gpt-4|Mistral-7B-Instruct-v0.2|...] bash ./scripts/run_re_tacrev.sh

ACE05

# To run this command, you shall firstly generate and save corresponding SLM's predictions.
# Then complete the env variables SLM_RES_PATH and OUT_PATH in the script.
MODEL_NAME=[gpt-4|Mistral-7B-Instruct-v0.2|...] bash ./scripts/run_ed_ace.sh

Citation

Please cite our paper if it is helpful for your work:

@misc{ma2023large,
      title={Large Language Model Is Not a Good Few-shot Information Extractor, but a Good Reranker for Hard Samples!}, 
      author={Yubo Ma and Yixin Cao and YongChing Hong and Aixin Sun},
      year={2023},
      eprint={2303.08559},
      archivePrefix={arXiv},
      primaryClass={cs.CL}
}