dfsbora / latex-math-model

0 stars 0 forks source link

Refine the Universal Transformer model to match the new LSTM model #15

Open ich-spiegelmann opened 2 months ago

ich-spiegelmann commented 2 months ago

Now, that the LSTM has finally achieved the level of the baseline from Karpathy's blog (http://karpathy.github.io/2015/05/21/rnn-effectiveness), the UT model also needs to be improved to match it.

As an option, it would be interesting to try the standard Trainer from the Hugging Face and a specialized Latex Tokenizer like e.g. MathBerta.