continuedev / continue

⏩ Continue is the leading open-source AI code assistant. You can connect any models and any context to build custom autocomplete and chat experiences inside VS Code and JetBrains
https://docs.continue.dev/
Apache License 2.0
19.22k stars 1.65k forks source link

Indexing errors on JetBrains #2320

Open Patrick-Erichsen opened 1 month ago

Patrick-Erichsen commented 1 month ago

Before submitting your bug report

Relevant environment info

- OS:
- Continue:
- IDE:
- Model:
- config.json:

Description

Multiple JetBrains users are reporting indexing issues.

Error: SQLITE_CONSTRAINT: UNIQUE constraint failed: tag_catalog.dir, tag_catalog.branch, tag_catalog.artifactId, tag_catalog.path, tag_catalog.cacheKey

Error: Failed to generate embeddings for 13 chunks with provider: OllamaEmbeddingsProvider::nomic-embed-text: Error: Failed to embed chunk:

Error: Failed to generate embeddings for 233 chunks with provider: OpenAIEmbeddingsProvider: BAAI/bge-m3: Error: {"error":{"message":"The parameter is invalid. Please check again. (request id: 2024091810405578781328035296047)","type":"upstream_error","param":"400","code":"bad_response_status_code"}}

Client network socket disconnected before secure TLS connection was established

tmibkr commented 1 month ago

IntelliJ Ultimate 2024.1 with Continue 0.0.64 or 0.0.69 (EAP) both result in this indexing error on Windows 10 Enterprise:

[2024-09-23T18:35:51] Indexing: 0.0% complete, elapsed time: 2s, 0.00 file/sec
[2024-09-23T18:35:51] Indexing: 25.0% complete, elapsed time: 2s, 48.50 file/sec
[2024-09-23T18:35:52] Indexing: 25.0% complete, elapsed time: 3s, 32.33 file/sec
[2024-09-23T18:36:12] error when indexing:  Error: Failed to generate embeddings for 1631 chunks with provider: _TransformersJsEmbeddingsProvider::all-MiniLM-L6-v2: Error: `local_files_only=true` or `env.allowRemoteModels=false` and file was not found locally at "C:\snapshot\continue\binary\models/all-MiniLM-L6-v2/tokenizer.json".
[2024-09-23T18:36:12] Indexing failed with error:  Error: Failed to generate embeddings for 1631 chunks with provider: _TransformersJsEmbeddingsProvider::all-MiniLM-L6-v2: Error: `local_files_only=true` or `env.allowRemoteModels=false` and file was not found locally at "C:\snapshot\continue\binary\models/all-MiniLM-L6-v2/tokenizer.json". LanceDbIndex.getEmbeddings, process.processTicksAndRejections, async LanceDbIndex.computeRows, async LanceDbIndex.update, async CodebaseIndexer.indexFiles, async CodebaseIndexer.refresh, async Core.refreshCodebaseIndex, async C:\snapshot\continue\binary\out\index.js:498256:7, async C:\snapshot\continue\binary\out\index.js:498574:28 

This path doesn't exist: C:\snapshot\continue . Maybe it's supposed to be a relative path from somewhere and is reverting to C: as a root instead?

liliwen365 commented 1 month ago

I also encountered this error in PyCharm. What impact does this have? Is there a solution available now? Thank you!
image

tmibkr commented 1 month ago

With IntelliJ Ultimate 2024.1 on Windows and Continue 0.0.75, re-indexing doesn't show an error in the plugin panel view any more, but it does log errors to core.log and no indexes turn up in $USERPROFILE/.continue/index/lancedb, and I get this error (see screenshot) when trying to use the @codebase feature. The listed path does not exist (C:\snapshot\continue\binary\models/all-MiniLM-L6-v2/tokenizer.json) and for some reason it has both \ and / for path separators. image

Searching through the codebase (https://github.com/search?q=repo%3Acontinuedev%2Fcontinue%20all-MiniLM-L6-v2&type=code), it seems like the embedding model is only included in the vscode extension and not the jetbrains one?

who52023 commented 3 weeks ago

With IntelliJ Ultimate 2024.1 on Windows and Continue 0.0.75, re-indexing doesn't show an error in the plugin panel view any more, but it does log errors to core.log and no indexes turn up in $USERPROFILE/.continue/index/lancedb, and I get this error (see screenshot) when trying to use the @codebase feature. The listed path does not exist (C:\snapshot\continue\binary\models/all-MiniLM-L6-v2/tokenizer.json) and for some reason it has both \ and / for path separators. image

Searching through the codebase (https://github.com/search?q=repo%3Acontinuedev%2Fcontinue%20all-MiniLM-L6-v2&type=code), it seems like the embedding model is only included in the vscode extension and not the jetbrains one?

i have the same qustion , i have no solution