-
### Describe the bug
I registered a custom model named test-llama-3-instruct which was already present in my directory using xinference register. After registration when the model is launched using…
-
### Question Validation
- [x] I have searched both the documentation and discord for an answer.
### Question
from llama_index.core.indices.property_graph import PropertyGraphIndex
from llama_index…
abc-w updated
1 month ago
-
Hi,
I am using nominc embedding model witihn a custom RAG solution
Here is the code
```
embed_model = SentenceTransformer("nomic-ai/nomic-embed-text-v1", trust_remote_code=True)
Settings.embe…
-
**Describe the bug**
**To Reproduce**
Steps to reproduce the behavior:
excute this command:
CMAKE_ARGS="-DLLAMA_CUDA=on -DLLAMA_NATIVE=off" pip install 'instructlab[cuda]'
and compile err…
-
Dear @dusty-nv ,
I'm trying the example code on web page: [Function Calling](https://dusty-nv.github.io/NanoLLM/chat.html#function-calling).
I tried both Llama-2-7b-chat-hf and Meta-Llama-3-8B-…
-
### What is the issue?
Upon running "ollama run gemma:2b" (though this happens for all tested models: llama3, phi, tinyllama), the loading animation appears and after ~5 minutes (estimate, untimed)…
-
With llama 3.1 has been introduced to help support multi-turn reasoning.
End of message. A message represents a possible stopping point for execution where the model can inform the executor that a…
-
I just follw the step, but when I run the following code :
# Load model directly
from transformers import AutoModel
model = AutoModel.from_pretrained("Efficient-Large-Model/Llama-3-VILA1.5-8B")
…
-
### Your current environment
```text
The output of `python collect_env.py`
Collecting environment information...
PyTorch version: 2.3.1+cu121
Is debug build: False
CUDA used to build PyTorch: 12…
-
#### Description
I encountered crashes in my application when attempting to load the `gemma-2b-it.gguf` and `Phi-3-mini-4k-instruct-q4.gguf` models. Below are the error messages and details for eac…