Closed jorgeantonio21 closed 3 months ago
This issue is related to #51
Does this still happen if you pull main branch? I believe for others this has been fixed. It may be the same issue with M3 that llama is facing
I'm fairly certian the problem is the softmax kernel producing inf on your machine, which makes the logits come out NaN, and triggers the blank token to be outputted, which is why you see no output at all. I will be revisiting the softmax kernel today or tomorrow to fix this
I pulled the main branch right now, and the problem persists.
Thank you so much @jafioti !
yes comment SoftmaxCompiler in luminal_metal lib.rs and Phi (and Llama) example will work on M3
@mikeseven Does it give proper outputs? In the other issue you mentioned it gives bad outputs
Sorry for the confusion. I wanted to say that the output looks correct but not as good as with llama. It looks to me a model accuracy issue.
Ok I'll close this for now then, thanks
Currently, I can't extract an output by running the phi3 example: