-
In the `gpt4_game_top30k_results.json` file, there are 20067 attacker win samples and 3287 defender win samples, with att/def ~ 6.1
While after SFTed the model using
```
torchrun --nproc_per_node…
thwu1 updated
4 months ago
-
It should be a minor extension
It shares gemma and llama features
And also a powerful model by itself
MMLU 71
[paper](https://storage.googleapis.com/deepmind-media/gemma/gemma-2-report.pdf)
…
-
- [ ] [Best way to add knowledge to a llm : r/LocalLLaMA](https://www.reddit.com/r/LocalLLaMA/comments/1ao2bzu/best_way_to_add_knowledge_to_a_llm/)
# Best way to add knowledge to a LLM: r/LocalLLaMA…
-
Steps to recreate:
Launch mmlu on an instance with multiple gpus. Run:
ilab model evaluate --model models/instructlab/granite-7b-lab --benchmark mmlu
Only 1 gpu is consumed. Adjusting batch…
-
-
# Proposed Feature
Add an efficient interface for generation probabilities on fixed prompt and completion pairs. For example:
```python
# ... load LLM or engine
prompt_completion_pairs = [
…
-
## model info
* base-model : baichuan-7b
* base-context-size : 4096
Did this phenomenon oberserved in your experiments?
In short context-window: Ntk > Yarn
![image](https://github.com/jquesn…
-
### 是否已有关于该错误的issue或讨论? | Is there an existing issue / discussion for this?
- [X] 我已经搜索过已有的issues和讨论 | I have searched the existing issues / discussions
### 该问题是否在FAQ中有解答? | Is there an existing ans…
-
Currently an MMLU-Run has the following output.
```
# KNOWLEDGE EVALUATION REPORT
## BASE MODEL
/home/ec2-user/.cache/instructlab/models/instructlab/granite-7b-lab
## MODEL
/home/ec2-user/…
-
## 🐛 Bug
Hello team,
Thanks for creating such an amazing engine. I ran Llama-3-8B-Instruct-q4f16_1-MLC in server mode with different batch sizes (2-128) but I still see my requests are being run …