huggingface / transformers

🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.
https://huggingface.co/transformers
Apache License 2.0
131.73k stars 26.23k forks source link

Siamese Multi-depth Transformer-based Hierarchical Encoder #9526

Open lalitpagaria opened 3 years ago

lalitpagaria commented 3 years ago

🌟 New model addition

Model description

Recently Google is published paper titled "Beyond 512 Tokens: Siamese Multi-depth Transformer-based Hierarchical Encoder for Long-Form Document Matching". And according to paper for long-form document matching SMITH model outperforms the previous state-of-the-art models including hierarchical attention, multi-depth attention-based hierarchical recurrent neural network, and BERT.

I feel it is will add value to already awesome transformers models collection :slightly_smiling_face:

Open source status

lalitpagaria commented 3 years ago

Linking Haystack issue https://github.com/deepset-ai/haystack/issues/719

Engineering-Geek commented 3 years ago

Frequent user of hugging face here, I'm a fan of this new publication and would love to see it implemented. Commenting here for the GitHub algorithm to ++

ChanCheeKean commented 3 years ago

Hi all, rather than waiting for the implementation in huggingface. Is there a simple way to utilize the pretrained model from the smith repo on our own dataset (to generate document embedding)?