-
How do I run a gguf multi-part model on ollaama?
https://huggingface.co/Qwen/Qwen2.5-7B-Instruct-GGUF/blob/main/qwen2.5-7b-instruct-fp16-00004-of-00004.gguf
qwen2.5-7b-instruct-fp16-00001-of-0000…
-
When loading the model [unsloth/Meta-Llama-3.1-8B-Instruct-bnb-4bit](https://huggingface.co/unsloth/Meta-Llama-3.1-8B-Instruct-bnb-4bit), I have the error : `AssertionError: model.safetensors.index.…
-
Hi neggles,
Thanks for this great work. Do you have any code/implementations for loading model from local safetensors?
-
Can we use safetensor models on this or only diffusers ?
-
Hello, I'm getting an incorrect hash after downloading the model
```
Traceback (most recent call last):
File "~/repos/training_results_v4.0/NVIDIA/benchmarks/llama2_70b_lora/implementations/nem…
-
Hellow
I am attempting to extract token embeddings from a model using the wordllama library. Although I have verified that the directory path is correct, I am encountering a FileNotFoundError. Below…
-
load dreambooth model from models/DreamBooth_LoRA/realisticVisionV60B1_v51VAE.safetensors
Traceback (most recent call last):
File "/home/tione/notebook/lskong2/softwares1/MotionClone/i2v_video_sam…
-
## Describe the bug
I ran the following, and it downloaded both uqff as well as full weights from Meta. I tried to skip -m, but it seems -m is required.
`./mistralrs-server -i vision-plain -m meta…
-
Confuzu, thank you for creating this useful tool. I wanted to make you aware of a possible issue.
I'm seeing that models and/or model versions are being skipped by the model grabber.
Prior to t…
-
I want something like this:
```py
import mlx.core as mx
import mlx.nn as nn
mx.metal.set_cache_limit(0)
def dload(x, *args, **kwargs):
la = mx.load(*args, **kwargs)
@mx.custom_function
…