-
llama stack build
Enter value for name (required): ollama
Enter value for distribution (default: local) (required): local-ollama
Enter value for api_providers (optional):
Enter value for image_ty…
-
### Prerequisites
- [X] I am running the latest code. Mention the version if possible as well.
- [X] I carefully followed the [README.md](https://github.com/ggerganov/llama.cpp/blob/master/README.md)…
-
### Description of the bug:
I tried running the example.py script given for quantization example, but for Llama. Wherever the reference to Gemma was made, I made appropriate references to Llama. The…
-
This issue serves to track performance on Metal hardware versus MLX and llama.cpp.
-
Here is a trace from my Intel Arc A770 via Docker:
```
$ ollama run deepseek-coder-v2
>>> write fizzbuzz
"""""""""""""""""""""""""""""""
```
And here is an trace from Arch linux running on …
-
### Description
When trying to run Unsloth fine-tuning script, encountering a Triton compilation error related to ReduceOpToLLVM.cpp.
### Error Message
```
python /data/ephemeral/home/unsloth_ex…
-
This bug popped up post `October-2024` tag. A glob file pattern makes it into HF's repo ID validator which causes it to fail.
Example:
```
model, tokenizer = FastLanguageModel.from_pretrain…
-
Thanks for adding VLM support to textgrad.
This doc describe how to use textgrad to do the autoprompt for [`gpt-4o`.](https://github.com/zou-group/textgrad/blob/main/examples/notebooks/Tutorial-Mul…
-
Currently the exporter does not handle higher order ops. Autocasts are expressed as HOPs:
```python
# Inside the ExportedProgram
...
class submod_1(torch.nn.Module):
def forward(self, e…
-
# Prerequisites
- [ x] I am running the latest code. Development is very rapid so there are no tagged versions as of now.
- [ x] I carefully followed the [README.md](https://github.com/abetlen/lla…