Open kmheckel opened 7 months ago
Isn't LRA with sequential CIFAR-10 instead of MNIST?
You're probably right - I didn't copy down all of the benchmarks and was basing the list off what was in the S5 repo so this may not be a perfect list. Will fix it.
On Wed, May 1, 2024, 22:45 Steven Abreu @.***> wrote:
Isn't LRA with sequential CIFAR-10 instead of MNIST?
— Reply to this email directly, view it on GitHub https://github.com/kmheckel/NeuroSSMs/issues/9#issuecomment-2089186568, or unsubscribe https://github.com/notifications/unsubscribe-auth/AMG7YKBFQEIRUXT527CFPMTZAFPBDAVCNFSM6AAAAABHCGBJQWVHI2DSMVQWIX3LMV43OSLTON2WKQ3PNVWWK3TUHMZDAOBZGE4DMNJWHA . You are receiving this because you authored the thread.Message ID: @.***>
We could also just do both (since MNIST is relatively fast)
Retrieval
Pathfinder
Path-X
General approach:
FP training as a reference baseline / target
PTQ as a baseline for quantization
QAT results
(optional) QAT starting from the FP checkpoint (i.e., finetuning)
[x] Sequential MNIST @stevenabreu7
LRA:
Bonus: