Closed saraswatpuneet closed 9 months ago
@saraswatpuneet this is a starting point and we build on these lines. The classes that inherits these classes will do the meat of the processing
Yep similar, have some basic abstract methods defined, we can add more as need arises, at minimum, llm should be able to generate entities and relationships, chat and more if we find anything
I like that everyone is on board :)
Designing base for LLM few things to keep in mind what faculties it will be used for Input data, ideally we can stream the data from ingestors into or how, do we need tokens or we can do a streaming one pass training on LLM as data is getting ingested ? Debanjan Datta and Nishant Gupta would know better as this is more on the research side for me Query interface : simple chat ? Some functions that can be used to communicate with their data with a chat interface Recommendations for entities and relationships Can we leverage vector space as a lens to LLM that can adjust it's weights accordingly ( I might be sounding crazy here )
So from a querent perspective our llms package must have such functionalities and possibly some good techniques we can find
Collectors->Injestors->Tokenizer->LLM:
We can have 1 jester class for all kinds of collector class. The Tokenizer is best kept separate for single and multi modal encoders. Why do you think?
I like it , I can imagine the flow of data through them in sequences which is awesome Yes I agree we should keep a separate implementation for each ;👍
Cool yeah propose a PR ideally add some readme(s) that could be a good starting point while the db part etc comes up (in progress)
Something like this