aimagelab / novelty-detection

Latent space autoregression for novelty detection.
MIT License
196 stars 60 forks source link

Details about training #2

Closed GinGinWang closed 5 years ago

GinGinWang commented 5 years ago

Hi, I am replicating your experiment's results. Could you give more the following details about the training on mnist and cifar10 which will replicate the results on your paper?

Thanks a lot! Best, Gin

mathue commented 5 years ago

Hi Gin,

from the supplementary material of the paper (chapter 7), one can infer lambda = 1 batch size = 256 learning rate = 10^-4 Optimizer = adam They stopped their training after a fixed number of 200 epochs (see closed issue) and monitored the LSA-loss (reconstruction loss plus autoregression loss (neg. llk)). However, if I train with these parameters, I only achieve an average AUROC-NS of 0.817 and 0.916 with a batch size of 64. Thus, they probably change some of these parameters for a better convergence. With batch size 256 it does not converge well but maybe I made a mistake. I attached my code for training (I changed the val() function in the mnist.py code to extract the first 90 % of the training data instead of the last 10 %). train.py.zip

DavideA commented 5 years ago

Hi and thank you for your interest.

@GinGinWang as correctly pointed out by @mathue, training details are provided in the supplementary material.

@mathue at a first glance, I can see a major issue in your training code:

Hope this helps, D

mathue commented 5 years ago

Hi Davide,

thanks alot for your reply. You are right, that's a shame, I made a mistake and trained on the 10 % validation set. When training on the 90 % training set for 200 epochs, I also achieved 0.972 AUROC-NS which is very close to what you reported in the paper.

Best regards Martin.

GinGinWang commented 5 years ago

Hi Davide, Sorry to bother you again. I replicate your experiments on cifar10 using all the parameters you recommended in your paper. After 200 epochs, I got AUROC 0.6. I also tried more epochs (1000 epochs), the AUROC did not increase. But I could have good results on the mnist dataset ... Is there any other training trick that is not mentioned in your paper? Thanks a lot for your help!