-
Congratulations on your excellent work! I attempted to run `bash scripts/run_streamingllm_lm.sh` to reproduce the results of streaming_llm, but I encountered the following error:
```
TypeError: ll…
-
Hey,
This was a really interesting solution to the KV cache for long context.
https://github.com/mit-han-lab/streaming-llm
I was wondering it could be implemented here. From the looks of thing…
-
Hi,
I was reading your paper and have a question about the "critical tokens". In Quest, the criticality of the tokens can change with different query tokens. While in StreamingLLM, the initial keys…
-
When I experiment with streamingllm in llama with this (https://github.com/NVIDIA/TensorRT-LLM/tree/main/examples/llama#run-llama-with-streamingllm), i wonder why always report length-related errors。w…
-
I run your instructions on the openbookqa task and got the following results:
full cache / dense:
`"openbookqa": {
"acc": 0.414,
"acc_stderr": 0.02204949796982787,
"acc_norm": 0…
-
Came across [this paper](https://github.com/mit-han-lab/streaming-llm/tree/main) on streaming LLMs that may be good for improving the efficiency of streaming.
-
### Describe the issue
_No response_
-
看论文主要是跟微调方法(如Positional Interpolation、NTK-Aware Scale ROPE和StreamingLLM)比较
有没有跟现有商业长上下文模型准确度对比评测结果?
想知道该技术方案的效果
-
From Section 3.2 in the paper:
```
When determining the relative distance and adding positional information to tokens, StreamingLLM
focuses on positions within the cache rather than those in the …
-
- [ ] [RichardAragon/MultiAgentLLM](https://github.com/richardaragon/multiagentllm)
# RichardAragon/MultiAgentLLM
**DESCRIPTION:** "Multi Agent Language Learning Machine (Multi Agent LLM)
(Update)…