-
### Is there an existing issue for this bug?
- [X] I have searched the existing issues
### 🐛 Describe the bug
I failed to run ChatGLM model with ColossalAI 0.3.6.
backtrace is here
----…
-
https://github.com/bytedance/ByteMLPerf/blob/main/byte_infer_perf/llm_perf/workloads/chatglm2-torch-fp16-6b.json
We run on A100-40G to get output logits with the below configuration:
```json
{
…
-
### 软件环境
```Markdown
- paddlepaddle-gpu: 0.0.0.post118
- paddlenlp: 2.8.0.post0
```
### 重复问题
- [X] I have searched the existing issues
### 错误描述
```Markdown
硬件:ai studio V100 32G
…
-
老哥,执行了sh tran.sh pre_tran.py,修改成了单卡模式,为啥会出现下面这个错误,搜了一下好像是“模型或数据未正确移至相应设备:”。
Number of trainable parameters = 1,431,996,416
0%| …
-
Hello! My project compiled perfectly with pyinstaller, but the exe file was slow to start, so I started trying to compile with Nuitka, but I ran into problems.
I used nuitka to compile the project e…
-
### 请提出你的问题
- 前提条件:单机单卡已经跑通chatglm2的lora微调训练代码;llama的多卡pp并行训练已跑通。
- 问题场景:想进一步尝试单机多卡,设置/chatglm2/lora_argument.json配置文件中的 "pipeline_parallel_degree": 4,然后参照官网样例,启动命令行:`srun --gres=gpu:4 python3 -u …
-
### 请提出你的问题
报错如下
Error Message Summary:
----------------------
ResourceExhaustedError:
Out of memory error on GPU 0. Cannot allocate 428.000000MB memory on GPU 0, 79.153320GB memory has been a…
-
[](https://github.com/THUDM/ChatGLM2-6B)
ghost updated
3 months ago
-
1) python code:
[test.txt](https://github.com/intel-analytics/ipex-llm/files/14797657/test.txt)
2) pip requirement:
[requirements.txt](https://github.com/intel-analytics/ipex-llm/files/14797675/req…
-
### Is there an existing issue for this?
- [X] I have searched the existing issues
### What is your question?
The following error occurred when I upgraded the version to milvus 2.3.3-gpu and the py…