Closed Ignotus closed 5 years ago
we can either do 13-15 o clock or 16-18?
13-14 works for me
Qs and Answers
Q: Training will take fairly long and our laptop-GPUs are not really made for that. What’s the proposed solution for getting access to GPUs?
Q: Few-Shot seems to be doing pretty much what we want. What is different in our project?
Q: Should we adapt it to U-net?
Q: Can we use instance normalization with a U-net architecture?
Q: Are we using triple consistency loss? Or formulating our own loss using the three papers?
Q: In the paper you shared (Few-Shot), how exactly are the projected embeddings used for the generator (just checking we understood it correctly)
Q: For the embedder (few-shot) are landmarks taken into account and if so, how?
Q: Propose adversarial predicting of landmarks like in the paper fader networks (add a discriminator to enforce embedding invariance wrt pose and mimics). Is this a good idea at all?
Q: Can we use our implementation afterwards (by law)? if not, why not?
A: Which law? Why not?
Dear Minh, @Ignotus
I would like some clarification to some of your answers.
First, In question 3 you replied that U-net is a later step and then in question 5 you say to have the pix2pix pipeline first. But pix2pix uses a U-net. So that leaves me a bit confused.
Second, to question 6, we understand the idea of extracting the features and averaging, just not how we practically inject the embedding into the generator. As it seems to be injected at multiple locations (i.e. not as a network input) and we need to implement it after all.
Third, to question 7, What does that mean practically, do we recolour the pictures that are going into the embedder? Or are we adding three channels to the input feature map? Or something else?
Fourth, for the final question we meant to ask: "Are we allowed to personally use our implementation after the course/project is finished. Do we maintain the rights?"
Finally, at what location will we meet today?
When do you guys want to meet? I may have a meeting from 3pm to 4pm.