Closed jisungyoon closed 4 years ago
Maybe, I think that is from the hub nodes of networks, and I briefly take a look from real embedding. Given node i, I collected the most similar 100 nodes, and calculate the sum of neighbor's sum of exponential cosine similarity of each node on the most similar 100 nodes. And get an average of 100 values as y, Appearance on node i as x value.
with std
with standard error
I think this effect makes the gravity model holds, but I am not sure with this equation. What's your thought?
And this approximation only holds on window size with 1
Sorry, I can't follow... could you define all symbols and provide more explanations?
F_ij is a flow between node i and j, and M_i is the sum of flow from node i. Cos(i,j) means cosine similarity between embedding vector i and j.
I wrote down a condition when does the gravity rule holds with the objective function of Word2vec.
Or, do you have a time on zoom? It is a complicated problem to talk about here.
Yeah maybe tomorrow sometime. But how does the in- and out-vector work here?
Yeah maybe tomorrow sometime. But how does the in- and out-vector work here?
Actually, I assumed that w_in == w_out, because of dividing two matrix makes a calculation difficult. If embeddings are good enough w_in ~~ w_out. When can you have a meeting on zoom?
let_s_do_math.pdf please check this pdf, and let's discuss about this issue
I've read through it but I think I need more time to think about it. Could you upload the latex source code to the gitub? (make it a subfolder in the paper directory). I will flesh it out as I work through the math.
I've read through it but I think I need more time to think about it. Could you upload the latex source code to the gitub? (make it a subfolder in the paper directory). I will flesh it out as I work through the math.
Or can we use overleaf for this latex file?
Yeah, overleaf is good for me!
Either share a link or add me as a collaborator with dakota.s.murray@gmail.com
Since this has now been put into the draft, lets close this issue
I found some very weak clue that why our model works. I am not good at mathematics, tell your thoughts about this results @yy @murrayds
I think it is related to the objective function of the word2vec model but is a little bit tricky. Because, in the word2vec model, they increase dot products between words, not exactly cosine similarity. But with cosine similarity, these equations hold
Equations said that there is an inverse relation between M_i and sum of neighbor's sum of exponential cosine similarity