This repository contains source code for the TaBERT model, a pre-trained language model for learning joint representations of natural language utterances and (semi-)structured tables for semantic parsing. TaBERT is pre-trained on a massive corpus of 26M Web tables and their associated natural language context, and could be used as a drop-in replacement of a semantic parsers original encoder to compute representations for utterances and table schemas (columns).
Other
580
stars
63
forks
source link
Minimal working example for Question answering on dataset using pre-trained model? #10
Hi, Can you provide a minimal example on how can I use the pre-trained model on question answering for new CSV or table-based datasets?
Something like :
predict( dataset, query )