-
# 🚀 Feature request
I've looked into the paper titled "[EL-Attention: Memory Efficient Lossless Attention for Generation](https://arxiv.org/abs/2105.04779)".
It proposes a method for calculating att…
-
Great work!
I would like to be able to cite your paper in my own work. However, I have encountered some confusion while reading your code.
1. Equation 6 in the paper does not seem to appear in the …
z1358 updated
1 month ago
-
### by using DynamicCache llm don't need to re compute the previous prompt. it can re use previous prompt kv cache!
### In gemini it's called context caching gemini & in anthropic it's called prompt …
-
## タイトル: StegaVision:注意機構によるステガノグラフィの強化
## リンク: https://arxiv.org/abs/2411.05838
## 概要:
画像ステガノグラフィは、画像内に秘密情報を埋め込む技術です。深層学習の発展により、この分野は大きく進歩しましたが、既存の手法は画質、埋め込み容量、およびセキュリティのバランスに課題がありました。本論文では、エンコーダ…
-
Thank you for your work.
after reading your paper, I have a question.
In Feature Split (FS) of sec. 3.2.2 Efficient Transformer, I was confused with the difference between this FS and window-att…
-
Thank you very much for your great work !
I encountered a problem while reading the source code: what is the role of num_tokens?
I found the `num_tokens` parameter in the source code of `IPAttnPr…
-
### Discussed in https://github.com/wazuh/wazuh/discussions/26545
Hello
We started to roll out the wazuh-agent to our Debian 12 machines. But all of them stop collecting logs via journald, when…
-
Dear Authors,
I hope you find this message.
First of all, I would like to express my appreciation for your excellent research. It is truly remarkable to see the quality you achieved using only s…
-
http://preview.d2l.ai.s3-website-us-west-2.amazonaws.com/d2l-en/master/chapter_recurrent-modern/seq2seq.html
http://preview.d2l.ai.s3-website-us-west-2.amazonaws.com/d2l-en/master/chapter_attention-m…
-
mgrA a gene known to repress tet38 (conferring resistance to tetracycline). But, mgrA has tetracycline as one of the terms in its Classification. Does Classification terms for gene actually indicates …