Closed dolaameng closed 8 years ago
Hi @dolaameng Sorry for late, reply.
What the shape (and if possible sample value), when you pass a word/phrase to word_embeddings
HI @iamaaditya thanks for the reply! You were right about the word vectors and those were good insights!
Again thank you for your great work!
@dolaameng
Images are continuous, small changes does not change the image much and it is still recognisable as old image. (That is why generative adversarial networks work). And the vector space of image is large. Moreover the features are obtained after training a very deep (19 layers) network and thus they are agnostic to perturbations.
Words, are not continuous. You cannot take a word "apple" and do +0.0001 and expect the word to still remain apple. If you do that with the word embedding of apple, it will find itself in some high-dimensional space which is not recognisable for the word apple. This happens also because our vocabulary might be 10K (or atmsot 50K) but a real valued vector of size 300 represents a much much larger space. Think of this way, in the whole solar system, you only have few marbles that you recognize. That is way using wrong embeddings as they are trained from makes these systems useless.
@iamaaditya : Thanks. I really like your explanations on the granularity in text vector space and how it affects the result. I understand that one main reason for mean-subtraction on images is to avoid gradient-issues in learning phase. But I am really interested know who this affect the testing phase. To verify what you said, I did some simple tests.
I know these tests might be too preliminary to conclude anything, but I had fan exploring your model! Look forward to seeing more interesting stuff from you! Thank you!
@dolaameng I am glad you had fun with it. I have another repo where I put code on how you could train your model. https://github.com/iamaaditya/VQA_Keras
Other people have uploaded much better models and have more analysis, if you get so interested in VQA task. Email me if you do not find the relevant resources.
Thanks! I will definitely check it out!
Hi,
Thank you for sharing the demo. I was trying to repeat the experiment but came across unexpected results.
My python libraries:
when I run
python demo.py
, here is the resultThe only change I made to the code is line 63 in demo.py
I used the default vectors due to a bug in recent spacy version, which shouldnt change the result too much?
I also noticed that in pre-processing the images for VGG16, there is no mean subtraction like
Would that cause a difference?
Appreciate your help on this! Thank you very much!