Open ALLinLLM opened 6 months ago
Since there are 4 experts adaptors using lora SFT in the paper, the next question is why not try MOE like Mixtral-8X7B?
+1
Since there are 4 experts adaptors using lora SFT in the paper, the next question is why not try MOE like Mixtral-8X7B?