-
On this platform, Ollama was installed successfully but got following error when running:
ollama run codellama:7b-instruct
Illegal instruction (core dumped)
-
With the publication of codellama, it became possible to run LLM on a local machine using [ollama](https://github.com/jmorganca/ollama) or [llama.cpp](https://github.com/ggerganov/llama.cpp).
**How …
-
The title says it all.
Can the configuration be modified so that we can use local LLMs?
-
The "/api/generate" is not functioning and display 404 on the Windows version (not WSL), despite the Ollama server running and "/" being accessible. The same code works on the Ollama server on m…
-
The function calling feature within the "local_function_calling" tool is currently not functioning as expected. When attempting to use this feature, it does not produce the intended results or throws …
-
Will start off by saying a *BIG* thank you for all your packages.
Using Ollama.
I had to fork the repo and create a branch with rev sha which was working.
Which is `226f8f0d9091e1feb113623ab6…
-
- [ ] [blog/mteb.md at main · huggingface/blog](https://github.com/huggingface/blog/blob/main/mteb.md?plain=1)
# Title: blog/mteb.md at main · huggingface/blog
**Description:**
"---
title: "MTEB: …
-
I have a machine with a lot of old parts in it, including 8 P40s and 2 Xeon E5-2667v2 CPUs.
I build llama.cpp using:
cmake -DLLAMA_AVX2=off -DLLAMA_F16C=off -DLLAMA_CUBLAS=on -DLLAMA_CUDA_FORCE_MM…
samr7 updated
2 months ago
-
## 🚀 Feature
Hello, thank you for all the great work, I truly like this project! ✨
It would be great to incorporate Prompt Lookup Decoding to speed up autoregressive decoding in LLMs. The project …
-
I was doing an inference work using codellama-2-7B.
Here is my code:
```
inputs_ids = tokenizer(prompt, return_tensors="pt").input_ids.to(self.device)
generate_ids=model.generate(inputs_ids,max_…