-
Llamaindex's `SemanticSplitterNodeParser` can sometimes produce chunks that are too large for the embedding model. Unfortunately there is no max length option for the semantic chunking to avoid this i…
-
I want to this alongside the migration to Unstructured. I'll figure out how helpful the difference is between my current implementation and something like spaCy would be for say, a long speech in a .t…
-
### Question Validation
- [X] I have searched both the documentation and discord for an answer.
### Question
@dosu iam using LlamaParseJsonNodeParser, for parsing the documents,iam using 8196 conte…
-
**Proposal**: Assess critical variables affecting semantic search quality on legal corpus before implementing semantic search on CL.
**Key Variables**:
- Embedding model
- Chunking strategy
- I…
-
**What would you like to be added**:
I propose to add a Semantic Search feature that enhances the ability to search and retrieve documents semantically. This functionality could be beneficial f…
-
Currently we do a character/word based chunking that is very simple. We should enhance our chunking strategies to possibly include:
* Recursive Character Chunking
* Token Based Chunking
* Documen…
-
## User Story
As a **engineer**, I would like **improve RAG performance** so that **the retrieved documents and generated answer are relevant to user's search query**.
## Detailed Description
Semanti…
-
**Is your feature request related to a problem? Please describe.**
Currently the `DocumentSplitter` in Haystack is relatively basic and recently we have seen that semantic splitting has greatly gaine…
-
Hi Greg,
Thanks a lot for you work!
I want to share with more optimized version of your function `combine_sentences` from the [tutorial about text splitting](https://github.com/FullStackRetrieva…
-
**Feature Overview (aka. Goal Summary)**
_An elevator pitch (value statement) that describes the Feature clearly and concisely. Complete during New status._
Converting a document with mixed elemen…
ktam3 updated
2 weeks ago