-
https://huggingface.co/apple/OpenELM
Has models ranging from 270M to 3B parameters. Would love to see more support for small models, since I'm stuck with 4gb VRAM currently. Tinyllama can't fill ev…
-
-
tinyllama which finetuning by function calling
-
hi unslothai, i got different inference result when using unsloth, i'v tested qwen1.5-chat and tinyllama-chat and got same issue, generate by unsloth always get a bad result compare with transformers …
-
So it appears that guided generation returns the requested structure like json only if the model has an infinite number of tokens it can generate, but otherwise very often it fails to close the struct…
-
### Feature request
Recently, we have added the ability to load `gguf` files within [transformers](https://huggingface.co/docs/hub/en/gguf).
The goal was to offer the possibility to users …
-
这个系列做的很不错,动手from scratch的过程学到了很多,希望tinyllama能更新一下tokenzier部分
-
### Question
Hello - having an issue getting this code to run in the browser. Using `Xenova/TinyLlama-1.1B-Chat-v1.0` on `"@xenova/transformers": "^2.13.2"`
It runs perfectly in node.
```ts
…
-
## Problem Description
When attempting to run Optimum Benchmark in vLLM using tensor parallelism across multiple AMD GPUs (MI210), I encounter a timeout error from NCCL watchdog. However, the bench…
-
As far as I know, one can pretrain fine using TinyStories:
```python
litgpt/pretrain.py --data litgpt.data.TinyStories
```
Should we add this to the documentation?
Right now, we only have …
rasbt updated
5 months ago