schmidt-ju / crat-pred

Other
62 stars 11 forks source link

LSTM ill-posed? #3

Closed Cram3r95 closed 1 year ago

Cram3r95 commented 2 years ago

Hi! This is an interesting repository with some SOTA concepts. Nevertheless, why do you use the output instead of the hidden state in the encoder?

According to the documentation:

output, (hn, cn) = rnn(input, (h0, c0))

We usually need hn (hidden state), whilst the output is usually used after the decoder, but you have this:

imagen

Any reason?

schmidt-ju commented 2 years ago

Hey,

I am not completely sure I understood your question correctly. From my understanding it should not make a difference, whether you use the the last timestep of the output of the LSTM or the hidden state. For multi-layer LSTMs, just make sure to use the hidden state of the last layer.

Keep in mind that I am using the nn.LSTM and not the nn.LSTMCell. The whole "feeding back the hidden state" is already done inside nn.LSTM.

Julian

Cram3r95 commented 2 years ago

Sorry for the late reply. I am already using the hidden state with num_layers = 1.

How are you able to obtain these metrics without map information? It is literally impossible to model these sequences:

imagen

If you do not provide map information

schmidt-ju commented 2 years ago

Hey,

are these predictions from our model? Keep in mind that metrics given in the paper are averaged over the whole splits. So there might be scenarios, where the prediction is off (which inevitable in some map-reliant scenarios).

Julian

Cram3r95 commented 2 years ago

No, these metrics and plots are from my code, but without map information, it is impossible to model curves, which are the main reason for metric drops. Do you obtain your metrics without map info?

schmidt-ju commented 1 year ago

Hey,

for sure there are some scenarios that are hard without map information. I can once again confirm that all values in our paper are achieved without map information.

Julian