Closed MatthewJA closed 8 years ago
One other idea could be to train the CNN in an unsupervised way, e.g. a CNN autoencoder. This would allow us to train on all the training data without biasing the features.
I suggest training the CNN on ALL the data for now. Document this peeking in your report.
If there is time later in the project, we can consider the following (in order):
Sounds good. Warm start CNN sounds like it could be a really good approach to take.
The CNN autoencoder you linked looks straightforward, too. I'll add this to milestone C to reconsider then.
Let's revisit this some time, possibly tomorrow?
Radio patches (left) and convolutional autoencoder reconstructions of the patches (right).
Reconstruction a smidgin too smooth, but for our purposes, it looks great.
Great! I'll rerun it a few times to try and nail down a decent network topology — I'd prefer less features than this provides, so I'll probably add another convolutional layer and maybe a dense layer.
I think my boundary conditions break with more convolutional layers, so I'm going to see if I can find another implementation and use my newfound convolutional autoencoder knowledge to get it working on the data.
Before you go down the route of finding features, visualise the IR and radio images of the positive examples that are classified negative by your predictor. 5-10 image patches from:
And for comparison, look at 5-10 patches where the score is >5.
At the same time, show the flux values (all other non-image features).
Alright, I'll get that done. #140
If you train logistic regression on the expert labels (100% accurate), you recover 85% balanced accuracy. If you train logistic regression on the crowd majority labels (85% accurate), you recover 85% balanced accuracy, too. This seems interesting! Maybe there's a maximum we're hitting.
I wonder if nonlinear and/or convolutional features would help.
From messing around with the feature extraction step of the pipeline, I've found that the CNN training massively affects the final accuracy. This raises two points: