-
Hello, I saw in the paper
but in the code, we see at https://github.com/ByungKwanLee/MoAI/blob/a7728a8d1c8df27d3221708a4ca4366e271f51c8/moai/arch/expert_module.py#L143
how this saves compute …
-
### Describe the bug
我好像没有找到用internevo训练然后转换成对应的hf的脚本?请问有提供嘛?
### Environment
官方代码
### Other information
_No response_
-
这个项目支持internlm2模型吗?
-
### Motivation
As of **July 2024**, in the field of open-source LLMs limited to **[CJK](https://en.wikipedia.org/wiki/CJK_characters) native-supported** models, [GLM-4-9B-Chat](https://huggingface.co…
-
如何用ppl 跑internlm2-20b模型的性能,也是使用pplnn-build/tools/benchmark_llama么?
-
### What is the issue?
In previous versions, I set the context length of each of my models to the maximum value that could be fully loaded onto the GPU memory. However, after the update, I found that…
-
### 📚 The doc issue
feature里有提到:[2024/04] TurboMind latest upgrade boosts GQA, rocketing the [internlm2-20b](https://huggingface.co/internlm/internlm2-20b) model inference to 16+ RPS, about 1.8x fa…
-
Is Poppler required to be downloaded for this project?
-
### Checklist
- [X] 1. I have searched related issues but cannot get the expected help.
- [X] 2. The bug has not been fixed in the latest version.
### Describe the bug
``` shell
D:\AI_model>lmdepl…
-
### Prerequisite
- [X] I have searched [Issues](https://github.com/open-compass/opencompass/issues/) and [Discussions](https://github.com/open-compass/opencompass/discussions) but cannot get the expe…