-
test
-
### Before submitting your bug report
- [X] I believe this is a bug. I'll try to join the [Continue Discord](https://discord.gg/NWtdYexhMs) for questions
- [X] I'm not able to find an [open issue](ht…
-
(base) mikekg@mikekg-mbp torchchat % python3 torchchat.py generate llama3 --device cpu --compile
Downloading meta-llama/Meta-Llama-3-8B-Instruct from HuggingFace...
Converting meta-llama/Meta-Llam…
-
HI,
I installed two Llama models using "Ollama run" in the terminal. Those occupy a significant space in disk and I need to free space to install a different model.
I tried Ollama rm command, bu…
-
You can see this in my PR where I tried to add the 'alpca' chat template:
https://github.com/ggerganov/llama.cpp/pull/7383
Later in the PR chat I actully copied and pasted the:
```
} els…
-
I would like to skip downloading models (either working directly with ollama or with Groq), is it possible?
During the 1st launch, even if I provide the Groq key, it keeps installing the models?
l…
-
### System Info
LangChain version: 0.0.330
Windows 11
Python 3.11.3
SQLAlchemy version: 2.0.23
### Who can help?
_No response_
### Information
- [ ] The official example notebooks/scripts
- [X…
-
cc @junrushao
## 🚀 Feature
The latest version of TVM enables different multi-GPU sharding configurations to reuse the same param weight shards. So, if you want to compile a model via MLC that has…
-
- [ ] [blog/mteb.md at main · huggingface/blog](https://github.com/huggingface/blog/blob/main/mteb.md?plain=1)
# Title: blog/mteb.md at main · huggingface/blog
**Description:**
"---
title: "MTEB: …
-
- [ ] [LLaVA/README.md at main · haotian-liu/LLaVA](https://github.com/haotian-liu/LLaVA/blob/main/README.md?plain=1)
# LLaVA/README.md at main · haotian-liu/LLaVA
## 🌋 LLaVA: Large Language and Vi…