-
### Is your feature request related to a problem? Please describe.
GroupChat uses a nested conversation between two agents. Currently it does not utilise the recent TransformMessages capability nor…
-
### Checked other resources
- [X] I added a very descriptive title to this issue.
- [X] I searched the LangChain documentation with the integrated search.
- [X] I used the GitHub search to find a…
-
**Describe the bug**
A clear and concise description of what the bug is.
Hey Team, trying to quantize mistral 8*22b with W8A8 recipe and failed with two issues with different versions:
1)
`…
-
### Describe the issue
First of all, thank you for your great contributions.
I have a similar question to the [issue 146](https://github.com/microsoft/LLMLingua/issues/146), I cannot reproduce the…
-
### Library name
TinyChatEngine
### Library description
TinyChatEngine: On-Device LLM Inference Library
### Source repository URL
https://github.com/mit-han-lab/TinyChatEngine
### Project homepa…
-
## List
- tutorials
- [ ] #4 - @seochan99
- [ ] #5 - @seochan99
- [ ] #6 - @seochan99
- [ ] #17 - @bananana0118
- [ ] graph.mdx
- [ ] index.mdx
- [ ] llm_chain.mdx
- [ ]…
-
What I want to achieve basically is re-ranking and prompt compression, before adding the retrieved docs to the context.
I read that it could drastically improve RAG performance. I think right now t…
-
Sorry to raise the problem but give no systematic analysis
It may be about to take me more time on more complete investigation over the "compression" ability of LLM as many may be support "compressio…
-
# URL
- https://arxiv.org/abs/2403.09636
# Affiliations
- Piotr Nawrot, N/A
- Adrian Łańcucki, N/A
- Marcin Chochowski, N/A
- David Tarjan, N/A
- Edoardo M. Ponti, N/A
# Abstract
- Transfo…
-
As mentioned in the paper, key concepts might get omitted either corrupted by the compression, in a way that the GPT can't process the compressed prompt.
You mention also there is an approach to op…