-
### System Info
Python 3.9.2
transformers 4.30.2
### Who can help?
@gante
### Information
- [X] The official example scripts
- [X] My own modified scripts
### Tasks
- [X] An officially support…
-
## Description
Multi model endpoint deployment in sagemaker through DJL serving is supposed to be supported. Here is the related [AWS page](https://docs.aws.amazon.com/sagemaker/latest/dg/deploy-mode…
-
Get assertion error with https://github.com/kharvd/gpt-cli
```
Traceback (most recent call last):
File "/home/gaocegege/applications/miniconda3/envs/dev/bin/gpt", line 8, in
sys.exit(main…
-
## 一言でいうと
言語モデルをマルチタスクかつ多言語に対応させるための効率的な学習データを探索した研究。英語単体でも他言語の性能が向上する。他言語の様々なタスクを英語/多言語への機械翻訳プロンプトで解かせることでさらに性能を底上げできるが、英語量が減ると英語性能は落ちる。
### 論文リンク
https://arxiv.org/abs/2211.01786
### 著者/…
-
Hi there.
My Cloud Atlas project will be using Bloomz, and I plan to incentivize all of my future users to run Petals on their machines.
But the first entry is not flawless, even i have problem…
-
The paper says that it only need 350G VRAM to train 175B GPT3 with rank =4. Can you elaborate more about how this is done? Like, do you use Megraton-deepspeed?
In my experiment with bloom-3b, fint…
-
after I do "python3 convert-hf-to-ggml.py bigscience/bloomz-7b1 ./models ",there was a problem loading the model:
```
model_path = "/aidata/yh/BelleGroup_BELLE-7B-1M-fp16/" # You can modify the path…
-
**Name of the feature**
*In general, the feature you want added should be supported by HuggingFace's [transformers](https://github.com/huggingface/transformers) library:*
- *If requesting a **model…
-
* CodeCarbon version: 2.7.1
* Python version: 3.10.14
* Operating System: Linux
### Description
I'm trying to estimate the carbon impact of LLM using available models on hugging face. So far I'm…
-
### System Info
The bug as described below occurs locally on my system with the following specs, and on google colab (see below for reproducible example):
```shell
- System: Ubuntu 22.04.3 LTS
…