This plugin enables a ChatGPT powered chatbot specifically focused on question answering over your Obsidian vault notes(markdown files).
https://user-images.githubusercontent.com/480507/236815106-2c38d137-9345-498e-8531-46494239d19a.mp4
Settings -> Community Plugins -> Browse
and searching for brAIn
.brAIn: Ingest vault docs
command in the command palette or by pressing Ctrl/Cmd + P
.brAIn: Open chat
command in the command palette or by pressing Ctrl/Cmd + P
.If you have new notes that you want to be indexed by the brAIn just run again brAIn: Ingest vault docs
The plugin uses a brAIn docker container running the brAIn server to enable the chatbot functionality. The brAIn server reads in your Obsidian vault documents and uses the OpenAI GPT-3 API to enable the question answering functionality. Once the server is running, you can chat with the bot through the plugin's interface.
If you have any issues or feature requests, please open an issue on GitHub.
This plugin is licensed under the MIT License.
By defauls the embeddings as generated locally so your docs are not leaving your machine, in case you check the OpenAI Embeddings check in settings, The OpenAI API is used by brAIn to generate the vector store during ingestion and to answer each question asked in the chat, but its API usage policies guarantee to us that OpenAI will not use data submitted by customers via its API to train or improve their models and any data submitted via API will be deleted after 30 days.
If you are generating the embeddings locally it could consume a lot of time and hardware resources (depending of your marchines and the number of documents) In case you want to generate the embeddings in OpenAI during the ingestion (embedding generation) a big amount of notes in your vault can lead to high expenses (~ 1000 notes = 1$), monitor your account and set API key limits to avoid scares
brAIn may occasionally generate incorrect or irrelevant content based on the user's input. Additionally, it may encounter errors when calling the OpenAI API if the API key is invalid or if there are issues with the OpenAI API service.
The final idea is to create a complete offline chatbot using an LLM, but so far the hardware requirements for these models are too high.