I have read your paper but still have some questions:
inference code? there are only two evaluation codes in your project
the input is image data and yaw value, the output is new embedding, but how to remapping it to frontal face visually.
Is it possible to implement in large scale, not the interesting result in your paper? only sample dataset is given
Thank you very much, I think the module is similar as the classical STN network, Good job~
Inference is quite simple, you just need to forward the face images through the trained model, and use the output feature for face verification or identification.
For visualization, please refer to this repo #11
"implement in large scale" means implement our algorithm on large dataset? Actually, experiment results reported in our paper are all implemented on large scale dataset. But those datasets we used, such as MsCeleb-1M belong to sensetime and I have no right to make them public available. So I only release a sample dataset to make sure you could go through the whole pipeline successfully and have a rough idea of how the training data should be organized.
I have read your paper but still have some questions: