Mintplex-Labs / anything-llm

The all-in-one Desktop & Docker AI application with built-in RAG, AI agents, and more.
https://anythingllm.com
MIT License
24.03k stars 2.41k forks source link

Issues Importing a Data Sources, Questions about AnythingLLM API Access, Local AnythingLLM Setup via Docker + More #487

Closed innovateworld closed 8 months ago

innovateworld commented 9 months ago

This might need to be split into multiple separate Issues. If so please let me know.

Issues Relating to My Goals (and possibly other people's) with AnythingLLM

There might not be a workaround for these but I have 2 primary goals with using AnythingLLM that I'm unable to use it for and I wanted to mention some additional separate concerns.

Goal 1: Use AnythingLLM to assist in generating Code using latest frameworks beyond LLM cutoff dates. Here are my problems:

A) Importing Certain Repositories from GitHub

B1) Non-Supported File Types

B2) Unable to easily select multiple docs in My Documents to delete several at once.

Goal 2: Use the AnythingLLM API from other development tools to run my LLM queries programmatically with my own external system prompts that would override the AnythingLLM system prompt to interact with the LLM but still be able to use the embeddings in the VectorDB that AnythingLLM generated with my custom Documents in my Workspace.

Separate Issues from Goals:

Running Locally with Docker is confusing

To be clear I like this solution overall and I run it currently with as a LocalLLm using Ollama + Litellm but .... As a separate concern I misunderstood that http://host.docker.internal:xxxx was supposed to be input from my Litellm port inside the AnythingLLM > LM Preference > Local AI Base URL instead of trying to figure out how to set Docker to connect to that port in the docker-compose.yml file lol).

Oh and since I'm on a roll.... I wanted to mention this:

B1.2) JSON files seemed to also not be supported! Pretty big deal!

dhlsam commented 9 months ago

website Scraping doesn't work ,via Docker

Primary server listening on port 3001 -- Working URL https://baijiahao.baidu.com/s?id=1786347651675278442&wfr=spider&for=pc -- getPageContent failed! Error: Could not find Chrome (ver. 119.0.6045.105). This can occur if either

  1. you did not perform an installation before running the script (e.g. npx puppeteer browsers install chrome) or
  2. your cache path is incorrectly configured (which is: /root/.cache/puppeteer). For (2), check out our guide on configuring puppeteer at https://pptr.dev/guides/configuration. at ChromeLauncher.resolveExecutablePath (file:///app/collector/node_modules/puppeteer-core/lib/esm/puppeteer/node/ProductLauncher.js:262:27) at ChromeLauncher.executablePath (file:///app/collector/node_modules/puppeteer-core/lib/esm/puppeteer/node/ChromeLauncher.js:213:25) at ChromeLauncher.computeLaunchArguments (file:///app/collector/node_modules/puppeteer-core/lib/esm/puppeteer/node/ChromeLauncher.js:107:37) at async ChromeLauncher.launch (file:///app/collector/node_modules/puppeteer-core/lib/esm/puppeteer/node/ProductLauncher.js:53:28) at async PuppeteerWebBaseLoader._scrape (/app/collector/node_modules/langchain/dist/document_loaders/web/puppeteer.cjs:42:25) at async PuppeteerWebBaseLoader.load (/app/collector/node_modules/langchain/dist/document_loaders/web/puppeteer.cjs:74:22) at async getPageContent (/app/collector/processLink/convert/generic.js:57:18) at async scrapeGenericUrl (/app/collector/processLink/convert/generic.js:11:19) at async processLink (/app/collector/processLink/index.js:6:10) at async /app/collector/index.js:48:33 TypeError: Cannot read properties of null (reading 'length') at scrapeGenericUrl (/app/collector/processLink/convert/generic.js:13:16) at async processLink (/app/collector/processLink/index.js:6:10) at async /app/collector/index.js:48:33
timothycarambat commented 8 months ago

Closing this as stale, almost all of the items listed are either now their own issues, or have been resolved, or are no longer relevant with recent changes 👍