-
### What happened + What you expected to happen
I want to use an environment with an observation space of 2 dimensions with the new API stack but I'm unable to do so as the `_get_encoder_config` me…
-
@keithito In Tacotron2 paper https://arxiv.org/abs/1712.05884, the authors mention that "The concatenation of the LSTM output and the attention context vector is then projected through a linear transf…
-
**Is your feature request related to a problem? Please describe.**
Build a deep learning model to predict future movements in interest rates based on historical data and economic indicators.
**Des…
-
Compare the effectiveness of CNN with attention against LSTM with attention. Metrics for comparison could include:
* Overall ROC-AUC score
* Some qualitative measure of the returned attention weig…
-
## ❔Question
我初始化模型:
```
model = AutoModel('seq2seq', custom_model_params={
"rnn_type": "lstm",
"bi_direction": False,
"rnn_size": 64,
"dense_siz…
-
使用paddle2onnx对Attention_LSTM模型进行转换,发现部分算子不支持:
There's 3 ops are not supported yet
=========== sequence_softmax ===========
=========== lstm ===========
=========== sequence_pool ===========
请…
-
Hey John! Here's the curriculum that I've worked on in the past. It's a bit less focused on language models as a sole topic, and more on modern ML from a broad perspective.
- Essential Concepts of …
zmaas updated
2 months ago
-
Would you please add the reference for the implementation details of the attention layer?
-
I made some changes to asr model.
encoder -> unidirectional lstm and monotonic attention.
After around 10000 steps, I can see that the loss is around 0.1
```
I0106 06:35:36.474763 14065233845632…
-
![image](https://github.com/king-yyf/CMeKG_tools/assets/63271390/7147f29f-a0a6-4fc4-88b0-24af1e293d8b)
请问这是transformer版本库导致的问题吗