Open neel04 opened 3 years ago
Code is out now:
https://github.com/google-research/google-research/tree/master/charformer (please note the different url - compared to paper url)
An unofficial PyTorch implementation for Charformer https://github.com/lucidrains/charformer-pytorch
Thanks for the great work! Will charformer be supported in the near future?
Still not supported yet
🌟 New model addition
Model description
arXiv = https://arxiv.org/pdf/2106.12672.pdf (pre-print; under review)
In this paper, they introduce a soft gradient-based subword tokenization module (GBST) that automatically learns latent subword representations from characters in a data-driven fashion. More importantly, is the introduction of Charformer, a deep Transformer model that integrates GBST and operates on the byte level.
Open source status