Closed agitter closed 6 years ago
They use data augmentation (flipped/rotated images). I am a bit skeptical about the "no additional explicit chemistry knowledge" claim, because the engines to render such images contain a lot of domain knowledge (bond lengths, symbols, angles, overlap detection ...). Nevertheless, it's an interesting (and funny) study, which again highlights the power of conv nets to extract features from visual input.
Closed by #774
https://arxiv.org/abs/1706.06689
Upon a quick read, they don't seem to start with a pre-trained Inception model. That was a little surprising given how few training instances they have for some of the tasks. I'd have to look carefully to see if the evaluation strategies are directly comparable, but #538 may report better performance on Tox21.