-
```
python train.py --base_config korquad/bidaf
```
```
2021-04-24 16:41:17,090 (trainer.py:356): [INFO] - # Train Mode.
2021-04-24 16:41:17,519 (trainer.py:389): [INFO] - Start - Batch Loss:…
-
Traceback (most recent call last):
File "demo.py", line 271, in
test()
File "demo.py", line 178, in test
encoder = encoder.cuda()
File "/home/iie/.conda/envs/s2v/lib/python2.7/site…
-
Apologies if I'm missing something obvious,
but when you get the output for the bidirectional RNN:
```
Xf = self.forward.get_output(train)
Xb = self.backward.get_output(train)
Xb = Xb[::-…
-
### 📚 The doc issue
The code snippet in PyTorch docs for [`nn.RNN`](https://pytorch.org/docs/stable/generated/torch.nn.RNN.html#torch.nn.RNN) seems to have a mistake.
Inside the `forward` functio…
-
### 🐛 Describe the bug
import torch.nn as nn
import torch
device = torch.device('cuda' if torch.cuda.is_available() else 'cpu')
class CRNN(nn.Module):
def __init__(self, img_channel…
-
Hi,
As mentioned [here](https://github.com/tensorflow/tensorflow/issues/63362#issuecomment-2226996257), a drop in model results was observed in the new version of TensorFlow using the new Keras. I …
-
In the batch translation example, encoder is:
```
class EncoderRNN(nn.Module):
def __init__(self, input_size, hidden_size, n_layers=1, dropout=0.1):
super(EncoderRNN, self).__init__()
…
-
I get the following error when I run the IMDB example:
Traceback (most recent call last):
File "imdb_birnn.py", line 77, in
model.add(BatchNormalization((24 \* maxseqlen,)))
File "/home/dxqu…
-
双向RNN里边需要传递batch里边每一条数据的确切长度,不然backward的结果完全就是错误的,费了好长时间才发现这个问题,麻烦请及时更正。
tf.nn.bidirectional_dynamic_rnn(cell_fw, cell_bw, seq, seq_length, initial_state_fw,initial_state_bw)
-
In the bi-lstm example, use outputs[-1] as the output of bi-lstm. I think it's wrong to do so. Because the outputs of rnn.static_bidirectional_rnn is depth-concatenated, which means outputs of bw_lstm…