-
This repo use LLM model by calling API
But, I wanna use it like a import a library, such as easyocr, pororo etc..
Can I use BetterOCR and LLM model like that?
-
Do you have any suggestions for a total newb who would like to get this running in docker on their synology NAS? I can follow instructions but do not have the best grasp of how this all works.
-
''ipex-llm[cpp]==2.5.0b20240527 is consistent with [v0.1.34] of ollama.
Our current version is consistent with [v0.1.39] of ollama.''
Is it possible to update supported ollama version to 0.3.x?
-
### Checklist
- [X] 1. I have searched related issues but cannot get the expected help.
- [X] 2. The bug has not been fixed in the latest version.
- [x] 3. Please note that if the bug-related iss…
wlwqq updated
3 months ago
-
### Check for existing issues
- [X] Completed
### Describe the feature
zed currently offers support for a limited number of AI assistants, but many model providers offer OpenAI-compatible interface…
-
### Your current environment
vllm version: 0.6.3.post1
gpu type: Quadro RTX 4000
### Model Input Dumps
_No response_
### 🐛 Describe the bug
I try to use gamme series such as `google/gemma-2-27…
-
**Describe the bug**
When installing ollama from ollama.org it detects arm64, but all models including llama3.2 don't utilize the GPU/NPU. As result the laptop battery is drained within 1 hour by 45%…
-
### 🚀 The feature, motivation and pitch
FLCE needs special handling for the soft capping in gemma2: https://github.com/huggingface/transformers/blob/main/src/transformers/models/gemma2/modeling_gemma…
-
When I run code based on examples/lightrag_ollama_demo.py I get following error, see below.
This happens with` llm_model_name="gemma2:2b"`
```
python3.11/site-packages/tiktoken/model.py", line …
-
**Qwen2**
warning: not compiled with GPU offload support, --n-gpu-layers option will be ignored
warning: see main README.md for information on enabling GPU BLAS support
Log start
main: build = 2…