-
Hi, I'm fine tuning an LLM using Soft Prompt Tuning using DeepSpeed via Accelerate implicitly, using the `deepspeed` param in `TrainingArguments`.
And all goes well until after the first epoch, whe…
-
### System Info
```
Package Version
------------------------ ------------
**accelerate 0.23.0**
aiohttp 3.8.6
aiosignal 1.3.1
ap…
-
It would also be nice to have a binary for this on conda-forge. You can see my start at implementing a recipe for this at https://github.com/conda-forge/staged-recipes/pull/24687/files As the CI over …
-
I'm very interested in getting this working. I am trying to get Mistral running on a TPU-VMs at GCP.
Using your example command:
python3 -m examples.serving.causal-lm.llama-2-chat --repo_id='mist…
-
## 🐛 Bug
I've been trying to figure out how to compile TheBloke/zephyr-7B-alpha-AWQ but have been running into error after error. Some tools state that awq isn't valid, others state mistral isn't.
…
-
### Your current environment
```text
# python3 collect_env.py
Collecting environment information...
PyTorch version: N/A
Is debug build: N/A
CUDA used to build PyTorch: N/A
ROCM used to build P…
-
## The problem
unable to perform test against android, using chrome driver executable path
## Environment
* Appium version: 1.17.0
* Last Appium version that did not exhibit the issue (if app…
-
The code below throws an assertion error:
```
from transformers import AutoModelForCausalLM, AutoTokenizer, pipeline
#model_name_or_path = "TheBloke/Mixtral-8x7B-Instruct-v0.1-GPTQ"
model_name…
-
I am trying to install the requirements on an Apple M1 Pro in a Macbook Pro and getting errors for building three wheels, have absolutely no idea how to fix
-
## Environment info
- `transformers` version: 4.2.2
- Platform: aws/codebuild/amazonlinux2-x86_64-standard:3.0 AND Windows-10-10.0.17763-SP0
- Python version: 3.8.3 AND 3.8.7
- PyTorch version (GP…