torchmd / torchmd-net

Training neural network potentials
MIT License
320 stars 72 forks source link

Clarifications of the method #144

Closed AlexDuvalinho closed 6 months ago

AlexDuvalinho commented 1 year ago

Hello, after reading the paper, I had several questions regarding your approach. Thanks a lot in advance for taking the time to answer them.

Your embedding layer is more complex than usual: your initial node representation already seems to depend on its neighbour’s representation.

Graph construction: you use a smooth cutoff function and describe some benefits. You describe a Transformers but still use a cutoff value.

You say the feature vector are passed through a normalization layer.

An intermediate node embedding (y_i) utilising attention scores is created and impact final x_i and vi embeddings. This step weights a projection of each neighbor’s representation ~ $a{ij} (W \cdot RBF(d_{ij}) \cdot \vec{V}_j)$ by the attention score.

The equivariant message m_ij (component of sum to obtain w_i) is obtained by multiplying s_ij^2 (i.e. v_j scaled by RBF(d_ij)) by the directional info r_ij; then adding to it s_ij^1 (i.e. v_j scaled by RBF(d_ij)) re-multiplied by v_j.

Invariant representation involves the scalar product of the equivariant vector v_i, projected with matrix U1 by (U2 v_i).

PhilippThoelke commented 1 year ago

Hi Alex, thank you for your interest in the work.

embedding layer

smooth cutoff

normalization

intermediate node embedding

equivariant message

scalar product

Hope this helps, feel free to ask if you have further questions.