Open ehartford opened 1 year ago
As was suggested in https://github.com/ggerganov/llama.cpp/issues/1333
Create a basic inference example for MosaicML MPT-7B model
I'll see how much my replit branch (from #131) would have to be adjusted... I think the main difference is the tokenizer and qkv clamping.
Looks decent, I'll create one PR for both then.
As was suggested in https://github.com/ggerganov/llama.cpp/issues/1333
Create a basic inference example for MosaicML MPT-7B model