maxjcohen / transformer

Implementation of Transformer model (originally from Attention is All You Need) applied to Time Series.
https://timeseriestransformer.readthedocs.io/en/latest/
GNU General Public License v3.0
843 stars 166 forks source link

Performance Evaluation #8

Closed DanielAtKrypton closed 3 years ago

DanielAtKrypton commented 4 years ago

I think it is worthwhile to evaluate the performance of the transformer. I have an indication that it is performing slower when compared to LSTM as of now.

We’re improving the state of scalable GPU computing in Python.

-- Matthew Rocklin

This post covers Python, performance, and GPUs. It lays out the current status, and describes future work.

It might be worth to evaluate performance boost with these techniques.

maxjcohen commented 4 years ago

Hi, yes Transformer currently performs slower, on this implementation, with this dataset.