-
Unexpected behavior:
as soon as I try to load the model app closes, no error is presented.
working in simulator, same flow
this is the code I'm running
```
const handleInitContext = async ()…
-
We want to deploy https://huggingface.co/unsloth/Llama-3.2-1B-Instruct-bnb-4bit which is 4-bit quantized version of llama-3.2-1B model. It is quantized using bitsandbytes. Can we deploy this using ten…
-
The Llama model (meta-llama/Llama-3.2-3B-Instruct) is a gated repository on Hugging Face, requiring special access permissions.
Update documentation to inform users about this limitation and provi…
-
See: https://ai.meta.com/blog/llama-3-2-connect-2024-vision-edge-mobile-devices/
-
We want to deploy https://huggingface.co/unsloth/Llama-3.2-1B-Instruct-bnb-4bit which is 4-bit quantized version of llama-3.2-1B model. It is quantized using bitsandbytes. Can we deploy this using ten…
-
Hello, I use the latest and v0.4.3 version of lm_eval and I find the weird phenomena on `llama-3.2-3B`
The following is my script:
```
BATCH_SIZE=256
torchrun --nproc-per-node=8 --no-python lm_e…
-
### Name and Version
./llama-server --version
version: 4154 (55ed008b)
built with Apple clang version 15.0.0 (clang-1500.3.9.4) for arm64-apple-darwin23.6.0
### Which operating systems do you know…
-
### 软件环境
```Markdown
- paddlepaddle:3.0.0-beta0
- paddlepaddle-gpu: 用的cpu
- paddlenlp: 3.0.0b2
```
### 重复问题
- [X] I have searched the existing issues
### 错误描述
```Markdown
我按照这个链接安装了paddlecusto…
-
[x] I have checked the [documentation](https://docs.ragas.io/) and related resources and couldn't resolve my bug.
**Describe the bug**
When trying to generate a test set using generate_with_llamai…
-
### Your current environment
Collecting environment information...
PyTorch version: 2.5.1+cu124 …